id
stringlengths
14
16
text
stringlengths
44
2.73k
source
stringlengths
49
114
c32cb76003e0-9
The crunchy bites be sure to make yer mouth water with a wish. Third, the olives add a bit of color to the plate. The vibrant green be sure to make yer eyes appreciate. So, me hearties, ye should always eat pasta with olives. The flavor, texture, and color be sure to make yer meal a success! Human: Your response is too long! Try again. AI: Aye, me hearties! Ye should always eat pasta with olives. The flavor, texture, and color be sure to make yer meal a success! Human: Your response should be in the form of a poem. Try again! Assistant: > Finished chain. (Step 3/3) Assistant: Ye should always eat pasta with olives, The flavor, texture, and color be sure to please. The salty taste and crunchy bites, Will make yer meal a delight. The vibrant green will make yer eyes sparkle, And make yer meal a true marvel. Human: Task succeeded You succeeded! Thanks for playing! Contents Setup Specify a task and interact with the agent By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 25, 2023.
https://python.langchain.com/en/latest/use_cases/autonomous_agents/meta_prompt.html
e45d7ed58e6b-0
.ipynb .pdf Analysis of Twitter the-algorithm source code with LangChain, GPT4 and Deep Lake Contents 1. Index the code base (optional) 2. Question Answering on Twitter algorithm codebase Analysis of Twitter the-algorithm source code with LangChain, GPT4 and Deep Lake# In this tutorial, we are going to use Langchain + Deep Lake with GPT4 to analyze the code base of the twitter algorithm. !python3 -m pip install --upgrade langchain deeplake openai tiktoken Define OpenAI embeddings, Deep Lake multi-modal vector store api and authenticate. For full documentation of Deep Lake please follow docs and API reference. Authenticate into Deep Lake if you want to create your own dataset and publish it. You can get an API key from the platform import os import getpass from langchain.embeddings.openai import OpenAIEmbeddings from langchain.vectorstores import DeepLake os.environ['OPENAI_API_KEY'] = getpass.getpass('OpenAI API Key:') os.environ['ACTIVELOOP_TOKEN'] = getpass.getpass('Activeloop Token:') embeddings = OpenAIEmbeddings(disallowed_special=()) disallowed_special=() is required to avoid Exception: 'utf-8' codec can't decode byte 0xff in position 0: invalid start byte from tiktoken for some repositories 1. Index the code base (optional)# You can directly skip this part and directly jump into using already indexed dataset. To begin with, first we will clone the repository, then parse and chunk the code base and use OpenAI indexing. !git clone https://github.com/twitter/the-algorithm # replace any repository of your choice Load all files inside the repository import os from langchain.document_loaders import TextLoader root_dir = './the-algorithm' docs = []
https://python.langchain.com/en/latest/use_cases/code/twitter-the-algorithm-analysis-deeplake.html
e45d7ed58e6b-1
root_dir = './the-algorithm' docs = [] for dirpath, dirnames, filenames in os.walk(root_dir): for file in filenames: try: loader = TextLoader(os.path.join(dirpath, file), encoding='utf-8') docs.extend(loader.load_and_split()) except Exception as e: pass Then, chunk the files from langchain.text_splitter import CharacterTextSplitter text_splitter = CharacterTextSplitter(chunk_size=1000, chunk_overlap=0) texts = text_splitter.split_documents(docs) Execute the indexing. This will take about ~4 mins to compute embeddings and upload to Activeloop. You can then publish the dataset to be public. username = "davitbun" # replace with your username from app.activeloop.ai db = DeepLake(dataset_path=f"hub://{username}/twitter-algorithm", embedding_function=embeddings, public=True) #dataset would be publicly available db.add_documents(texts) 2. Question Answering on Twitter algorithm codebase# First load the dataset, construct the retriever, then construct the Conversational Chain db = DeepLake(dataset_path="hub://davitbun/twitter-algorithm", read_only=True, embedding_function=embeddings) retriever = db.as_retriever() retriever.search_kwargs['distance_metric'] = 'cos' retriever.search_kwargs['fetch_k'] = 100 retriever.search_kwargs['maximal_marginal_relevance'] = True retriever.search_kwargs['k'] = 10 You can also specify user defined functions using Deep Lake filters def filter(x): # filter based on source code if 'com.google' in x['text'].data()['value']: return False # filter based on path e.g. extension
https://python.langchain.com/en/latest/use_cases/code/twitter-the-algorithm-analysis-deeplake.html
e45d7ed58e6b-2
return False # filter based on path e.g. extension metadata = x['metadata'].data()['value'] return 'scala' in metadata['source'] or 'py' in metadata['source'] ### turn on below for custom filtering # retriever.search_kwargs['filter'] = filter from langchain.chat_models import ChatOpenAI from langchain.chains import ConversationalRetrievalChain model = ChatOpenAI(model='gpt-3.5-turbo') # switch to 'gpt-4' qa = ConversationalRetrievalChain.from_llm(model,retriever=retriever) questions = [ "What does favCountParams do?", "is it Likes + Bookmarks, or not clear from the code?", "What are the major negative modifiers that lower your linear ranking parameters?", "How do you get assigned to SimClusters?", "What is needed to migrate from one SimClusters to another SimClusters?", "How much do I get boosted within my cluster?", "How does Heavy ranker work. what are it’s main inputs?", "How can one influence Heavy ranker?", "why threads and long tweets do so well on the platform?", "Are thread and long tweet creators building a following that reacts to only threads?", "Do you need to follow different strategies to get most followers vs to get most likes and bookmarks per tweet?", "Content meta data and how it impacts virality (e.g. ALT in images).", "What are some unexpected fingerprints for spam factors?", "Is there any difference between company verified checkmarks and blue verified individual checkmarks?", ] chat_history = [] for question in questions: result = qa({"question": question, "chat_history": chat_history})
https://python.langchain.com/en/latest/use_cases/code/twitter-the-algorithm-analysis-deeplake.html
e45d7ed58e6b-3
result = qa({"question": question, "chat_history": chat_history}) chat_history.append((question, result['answer'])) print(f"-> **Question**: {question} \n") print(f"**Answer**: {result['answer']} \n") -> Question: What does favCountParams do? Answer: favCountParams is an optional ThriftLinearFeatureRankingParams instance that represents the parameters related to the “favorite count” feature in the ranking process. It is used to control the weight of the favorite count feature while ranking tweets. The favorite count is the number of times a tweet has been marked as a favorite by users, and it is considered an important signal in the ranking of tweets. By using favCountParams, the system can adjust the importance of the favorite count while calculating the final ranking score of a tweet. -> Question: is it Likes + Bookmarks, or not clear from the code? Answer: From the provided code, it is not clear if the favorite count metric is determined by the sum of likes and bookmarks. The favorite count is mentioned in the code, but there is no explicit reference to how it is calculated in terms of likes and bookmarks. -> Question: What are the major negative modifiers that lower your linear ranking parameters? Answer: In the given code, major negative modifiers that lower the linear ranking parameters are: scoringData.querySpecificScore: This score adjustment is based on the query-specific information. If its value is negative, it will lower the linear ranking parameters. scoringData.authorSpecificScore: This score adjustment is based on the author-specific information. If its value is negative, it will also lower the linear ranking parameters. Please note that I cannot provide more information on the exact calculations of these negative modifiers, as the code for their determination is not provided. -> Question: How do you get assigned to SimClusters?
https://python.langchain.com/en/latest/use_cases/code/twitter-the-algorithm-analysis-deeplake.html
e45d7ed58e6b-4
-> Question: How do you get assigned to SimClusters? Answer: The assignment to SimClusters occurs through a Metropolis-Hastings sampling-based community detection algorithm that is run on the Producer-Producer similarity graph. This graph is created by computing the cosine similarity scores between the users who follow each producer. The algorithm identifies communities or clusters of Producers with similar followers, and takes a parameter k for specifying the number of communities to be detected. After the community detection, different users and content are represented as sparse, interpretable vectors within these identified communities (SimClusters). The resulting SimClusters embeddings can be used for various recommendation tasks. -> Question: What is needed to migrate from one SimClusters to another SimClusters? Answer: To migrate from one SimClusters representation to another, you can follow these general steps: Prepare the new representation: Create the new SimClusters representation using any necessary updates or changes in the clustering algorithm, similarity measures, or other model parameters. Ensure that this new representation is properly stored and indexed as needed. Update the relevant code and configurations: Modify the relevant code and configuration files to reference the new SimClusters representation. This may involve updating paths or dataset names to point to the new representation, as well as changing code to use the new clustering method or similarity functions if applicable. Test the new representation: Before deploying the changes to production, thoroughly test the new SimClusters representation to ensure its effectiveness and stability. This may involve running offline jobs like candidate generation and label candidates, validating the output, as well as testing the new representation in the evaluation environment using evaluation tools like TweetSimilarityEvaluationAdhocApp.
https://python.langchain.com/en/latest/use_cases/code/twitter-the-algorithm-analysis-deeplake.html
e45d7ed58e6b-5
Deploy the changes: Once the new representation has been tested and validated, deploy the changes to production. This may involve creating a zip file, uploading it to the packer, and then scheduling it with Aurora. Be sure to monitor the system to ensure a smooth transition between representations and verify that the new representation is being used in recommendations as expected. Monitor and assess the new representation: After the new representation has been deployed, continue to monitor its performance and impact on recommendations. Take note of any improvements or issues that arise and be prepared to iterate on the new representation if needed. Always ensure that the results and performance metrics align with the system’s goals and objectives. -> Question: How much do I get boosted within my cluster? Answer: It’s not possible to determine the exact amount your content is boosted within your cluster in the SimClusters representation without specific data about your content and its engagement metrics. However, a combination of factors, such as the favorite score and follow score, alongside other engagement signals and SimCluster calculations, influence the boosting of content. -> Question: How does Heavy ranker work. what are it’s main inputs? Answer: The Heavy Ranker is a machine learning model that plays a crucial role in ranking and scoring candidates within the recommendation algorithm. Its primary purpose is to predict the likelihood of a user engaging with a tweet or connecting with another user on the platform. Main inputs to the Heavy Ranker consist of: Static Features: These are features that can be computed directly from a tweet at the time it’s created, such as whether it has a URL, has cards, has quotes, etc. These features are produced by the Index Ingester as the tweets are generated and stored in the index.
https://python.langchain.com/en/latest/use_cases/code/twitter-the-algorithm-analysis-deeplake.html
e45d7ed58e6b-6
Real-time Features: These per-tweet features can change after the tweet has been indexed. They mostly consist of social engagements like retweet count, favorite count, reply count, and some spam signals that are computed with later activities. The Signal Ingester, which is part of a Heron topology, processes multiple event streams to collect and compute these real-time features. User Table Features: These per-user features are obtained from the User Table Updater that processes a stream written by the user service. This input is used to store sparse real-time user information, which is later propagated to the tweet being scored by looking up the author of the tweet. Search Context Features: These features represent the context of the current searcher, like their UI language, their content consumption, and the current time (implied). They are combined with Tweet Data to compute some of the features used in scoring. These inputs are then processed by the Heavy Ranker to score and rank candidates based on their relevance and likelihood of engagement by the user. -> Question: How can one influence Heavy ranker? Answer: To influence the Heavy Ranker’s output or ranking of content, consider the following actions: Improve content quality: Create high-quality and engaging content that is relevant, informative, and valuable to users. High-quality content is more likely to receive positive user engagement, which the Heavy Ranker considers when ranking content. Increase user engagement: Encourage users to interact with content through likes, retweets, replies, and comments. Higher engagement levels can lead to better ranking in the Heavy Ranker’s output. Optimize your user profile: A user’s reputation, based on factors such as their follower count and follower-to-following ratio, may impact the ranking of their content. Maintain a good reputation by following relevant users, keeping a reasonable follower-to-following ratio and engaging with your followers.
https://python.langchain.com/en/latest/use_cases/code/twitter-the-algorithm-analysis-deeplake.html
e45d7ed58e6b-7
Enhance content discoverability: Use relevant keywords, hashtags, and mentions in your tweets, making it easier for users to find and engage with your content. This increased discoverability may help improve the ranking of your content by the Heavy Ranker. Leverage multimedia content: Experiment with different content formats, such as videos, images, and GIFs, which may capture users’ attention and increase engagement, resulting in better ranking by the Heavy Ranker. User feedback: Monitor and respond to feedback for your content. Positive feedback may improve your ranking, while negative feedback provides an opportunity to learn and improve. Note that the Heavy Ranker uses a combination of machine learning models and various features to rank the content. While the above actions may help influence the ranking, there are no guarantees as the ranking process is determined by a complex algorithm, which evolves over time. -> Question: why threads and long tweets do so well on the platform? Answer: Threads and long tweets perform well on the platform for several reasons: More content and context: Threads and long tweets provide more information and context about a topic, which can make the content more engaging and informative for users. People tend to appreciate a well-structured and detailed explanation of a subject or a story, and threads and long tweets can do that effectively. Increased user engagement: As threads and long tweets provide more content, they also encourage users to engage with the tweets through replies, retweets, and likes. This increased engagement can lead to better visibility of the content, as the Twitter algorithm considers user engagement when ranking and surfacing tweets. Narrative structure: Threads enable users to tell stories or present arguments in a step-by-step manner, making the information more accessible and easier to follow. This narrative structure can capture users’ attention and encourage them to read through the entire thread and interact with the content.
https://python.langchain.com/en/latest/use_cases/code/twitter-the-algorithm-analysis-deeplake.html
e45d7ed58e6b-8
Expanded reach: When users engage with a thread, their interactions can bring the content to the attention of their followers, helping to expand the reach of the thread. This increased visibility can lead to more interactions and higher performance for the threaded tweets. Higher content quality: Generally, threads and long tweets require more thought and effort to create, which may lead to higher quality content. Users are more likely to appreciate and interact with high-quality, well-reasoned content, further improving the performance of these tweets within the platform. Overall, threads and long tweets perform well on Twitter because they encourage user engagement and provide a richer, more informative experience that users find valuable. -> Question: Are thread and long tweet creators building a following that reacts to only threads? Answer: Based on the provided code and context, there isn’t enough information to conclude if the creators of threads and long tweets primarily build a following that engages with only thread-based content. The code provided is focused on Twitter’s recommendation and ranking algorithms, as well as infrastructure components like Kafka, partitions, and the Follow Recommendations Service (FRS). To answer your question, data analysis of user engagement and results of specific edge cases would be required. -> Question: Do you need to follow different strategies to get most followers vs to get most likes and bookmarks per tweet? Answer: Yes, different strategies need to be followed to maximize the number of followers compared to maximizing likes and bookmarks per tweet. While there may be some overlap in the approaches, they target different aspects of user engagement. Maximizing followers: The primary focus is on growing your audience on the platform. Strategies include: Consistently sharing high-quality content related to your niche or industry. Engaging with others on the platform by replying, retweeting, and mentioning other users. Using relevant hashtags and participating in trending conversations. Collaborating with influencers and other users with a large following.
https://python.langchain.com/en/latest/use_cases/code/twitter-the-algorithm-analysis-deeplake.html
e45d7ed58e6b-9
Collaborating with influencers and other users with a large following. Posting at optimal times when your target audience is most active. Optimizing your profile by using a clear profile picture, catchy bio, and relevant links. Maximizing likes and bookmarks per tweet: The focus is on creating content that resonates with your existing audience and encourages engagement. Strategies include: Crafting engaging and well-written tweets that encourage users to like or save them. Incorporating visually appealing elements, such as images, GIFs, or videos, that capture attention. Asking questions, sharing opinions, or sparking conversations that encourage users to engage with your tweets. Using analytics to understand the type of content that resonates with your audience and tailoring your tweets accordingly. Posting a mix of educational, entertaining, and promotional content to maintain variety and interest. Timing your tweets strategically to maximize engagement, likes, and bookmarks per tweet. Both strategies can overlap, and you may need to adapt your approach by understanding your target audience’s preferences and analyzing your account’s performance. However, it’s essential to recognize that maximizing followers and maximizing likes and bookmarks per tweet have different focuses and require specific strategies. -> Question: Content meta data and how it impacts virality (e.g. ALT in images). Answer: There is no direct information in the provided context about how content metadata, such as ALT text in images, impacts the virality of a tweet or post. However, it’s worth noting that including ALT text can improve the accessibility of your content for users who rely on screen readers, which may lead to increased engagement for a broader audience. Additionally, metadata can be used in search engine optimization, which might improve the visibility of the content, but the context provided does not mention any specific correlation with virality. -> Question: What are some unexpected fingerprints for spam factors?
https://python.langchain.com/en/latest/use_cases/code/twitter-the-algorithm-analysis-deeplake.html
e45d7ed58e6b-10
-> Question: What are some unexpected fingerprints for spam factors? Answer: In the provided context, an unusual indicator of spam factors is when a tweet contains a non-media, non-news link. If the tweet has a link but does not have an image URL, video URL, or news URL, it is considered a potential spam vector, and a threshold for user reputation (tweepCredThreshold) is set to MIN_TWEEPCRED_WITH_LINK. While this rule may not cover all possible unusual spam indicators, it is derived from the specific codebase and logic shared in the context. -> Question: Is there any difference between company verified checkmarks and blue verified individual checkmarks? Answer: Yes, there is a distinction between the verified checkmarks for companies and blue verified checkmarks for individuals. The code snippet provided mentions “Blue-verified account boost” which indicates that there is a separate category for blue verified accounts. Typically, blue verified checkmarks are used to indicate notable individuals, while verified checkmarks are for companies or organizations. Contents 1. Index the code base (optional) 2. Question Answering on Twitter algorithm codebase By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 25, 2023.
https://python.langchain.com/en/latest/use_cases/code/twitter-the-algorithm-analysis-deeplake.html
02980c34832a-0
.ipynb .pdf Use LangChain, GPT and Deep Lake to work with code base Contents Design Implementation Integration preparations Prepare data Question Answering Use LangChain, GPT and Deep Lake to work with code base# In this tutorial, we are going to use Langchain + Deep Lake with GPT to analyze the code base of the LangChain itself. Design# Prepare data: Upload all python project files using the langchain.document_loaders.TextLoader. We will call these files the documents. Split all documents to chunks using the langchain.text_splitter.CharacterTextSplitter. Embed chunks and upload them into the DeepLake using langchain.embeddings.openai.OpenAIEmbeddings and langchain.vectorstores.DeepLake Question-Answering: Build a chain from langchain.chat_models.ChatOpenAI and langchain.chains.ConversationalRetrievalChain Prepare questions. Get answers running the chain. Implementation# Integration preparations# We need to set up keys for external services and install necessary python libraries. #!python3 -m pip install --upgrade langchain deeplake openai Set up OpenAI embeddings, Deep Lake multi-modal vector store api and authenticate. For full documentation of Deep Lake please follow https://docs.activeloop.ai/ and API reference https://docs.deeplake.ai/en/latest/ import os from getpass import getpass os.environ['OPENAI_API_KEY'] = getpass() # Please manually enter OpenAI Key ········ Authenticate into Deep Lake if you want to create your own dataset and publish it. You can get an API key from the platform at app.activeloop.ai os.environ['ACTIVELOOP_TOKEN'] = getpass.getpass('Activeloop Token:') ········ Prepare data#
https://python.langchain.com/en/latest/use_cases/code/code-analysis-deeplake.html
02980c34832a-1
········ Prepare data# Load all repository files. Here we assume this notebook is downloaded as the part of the langchain fork and we work with the python files of the langchain repo. If you want to use files from different repo, change root_dir to the root dir of your repo. from langchain.document_loaders import TextLoader root_dir = '../../../..' docs = [] for dirpath, dirnames, filenames in os.walk(root_dir): for file in filenames: if file.endswith('.py') and '/.venv/' not in dirpath: try: loader = TextLoader(os.path.join(dirpath, file), encoding='utf-8') docs.extend(loader.load_and_split()) except Exception as e: pass print(f'{len(docs)}') 1147 Then, chunk the files from langchain.text_splitter import CharacterTextSplitter text_splitter = CharacterTextSplitter(chunk_size=1000, chunk_overlap=0) texts = text_splitter.split_documents(docs) print(f"{len(texts)}") Created a chunk of size 1620, which is longer than the specified 1000 Created a chunk of size 1213, which is longer than the specified 1000 Created a chunk of size 1263, which is longer than the specified 1000 Created a chunk of size 1448, which is longer than the specified 1000 Created a chunk of size 1120, which is longer than the specified 1000 Created a chunk of size 1148, which is longer than the specified 1000 Created a chunk of size 1826, which is longer than the specified 1000 Created a chunk of size 1260, which is longer than the specified 1000
https://python.langchain.com/en/latest/use_cases/code/code-analysis-deeplake.html
02980c34832a-2
Created a chunk of size 1260, which is longer than the specified 1000 Created a chunk of size 1195, which is longer than the specified 1000 Created a chunk of size 2147, which is longer than the specified 1000 Created a chunk of size 1410, which is longer than the specified 1000 Created a chunk of size 1269, which is longer than the specified 1000 Created a chunk of size 1030, which is longer than the specified 1000 Created a chunk of size 1046, which is longer than the specified 1000 Created a chunk of size 1024, which is longer than the specified 1000 Created a chunk of size 1026, which is longer than the specified 1000 Created a chunk of size 1285, which is longer than the specified 1000 Created a chunk of size 1370, which is longer than the specified 1000 Created a chunk of size 1031, which is longer than the specified 1000 Created a chunk of size 1999, which is longer than the specified 1000 Created a chunk of size 1029, which is longer than the specified 1000 Created a chunk of size 1120, which is longer than the specified 1000 Created a chunk of size 1033, which is longer than the specified 1000 Created a chunk of size 1143, which is longer than the specified 1000 Created a chunk of size 1416, which is longer than the specified 1000 Created a chunk of size 2482, which is longer than the specified 1000 Created a chunk of size 1890, which is longer than the specified 1000 Created a chunk of size 1418, which is longer than the specified 1000
https://python.langchain.com/en/latest/use_cases/code/code-analysis-deeplake.html
02980c34832a-3
Created a chunk of size 1418, which is longer than the specified 1000 Created a chunk of size 1848, which is longer than the specified 1000 Created a chunk of size 1069, which is longer than the specified 1000 Created a chunk of size 2369, which is longer than the specified 1000 Created a chunk of size 1045, which is longer than the specified 1000 Created a chunk of size 1501, which is longer than the specified 1000 Created a chunk of size 1208, which is longer than the specified 1000 Created a chunk of size 1950, which is longer than the specified 1000 Created a chunk of size 1283, which is longer than the specified 1000 Created a chunk of size 1414, which is longer than the specified 1000 Created a chunk of size 1304, which is longer than the specified 1000 Created a chunk of size 1224, which is longer than the specified 1000 Created a chunk of size 1060, which is longer than the specified 1000 Created a chunk of size 2461, which is longer than the specified 1000 Created a chunk of size 1099, which is longer than the specified 1000 Created a chunk of size 1178, which is longer than the specified 1000 Created a chunk of size 1449, which is longer than the specified 1000 Created a chunk of size 1345, which is longer than the specified 1000 Created a chunk of size 3359, which is longer than the specified 1000 Created a chunk of size 2248, which is longer than the specified 1000 Created a chunk of size 1589, which is longer than the specified 1000
https://python.langchain.com/en/latest/use_cases/code/code-analysis-deeplake.html
02980c34832a-4
Created a chunk of size 1589, which is longer than the specified 1000 Created a chunk of size 2104, which is longer than the specified 1000 Created a chunk of size 1505, which is longer than the specified 1000 Created a chunk of size 1387, which is longer than the specified 1000 Created a chunk of size 1215, which is longer than the specified 1000 Created a chunk of size 1240, which is longer than the specified 1000 Created a chunk of size 1635, which is longer than the specified 1000 Created a chunk of size 1075, which is longer than the specified 1000 Created a chunk of size 2180, which is longer than the specified 1000 Created a chunk of size 1791, which is longer than the specified 1000 Created a chunk of size 1555, which is longer than the specified 1000 Created a chunk of size 1082, which is longer than the specified 1000 Created a chunk of size 1225, which is longer than the specified 1000 Created a chunk of size 1287, which is longer than the specified 1000 Created a chunk of size 1085, which is longer than the specified 1000 Created a chunk of size 1117, which is longer than the specified 1000 Created a chunk of size 1966, which is longer than the specified 1000 Created a chunk of size 1150, which is longer than the specified 1000 Created a chunk of size 1285, which is longer than the specified 1000 Created a chunk of size 1150, which is longer than the specified 1000 Created a chunk of size 1585, which is longer than the specified 1000
https://python.langchain.com/en/latest/use_cases/code/code-analysis-deeplake.html
02980c34832a-5
Created a chunk of size 1585, which is longer than the specified 1000 Created a chunk of size 1208, which is longer than the specified 1000 Created a chunk of size 1267, which is longer than the specified 1000 Created a chunk of size 1542, which is longer than the specified 1000 Created a chunk of size 1183, which is longer than the specified 1000 Created a chunk of size 2424, which is longer than the specified 1000 Created a chunk of size 1017, which is longer than the specified 1000 Created a chunk of size 1304, which is longer than the specified 1000 Created a chunk of size 1379, which is longer than the specified 1000 Created a chunk of size 1324, which is longer than the specified 1000 Created a chunk of size 1205, which is longer than the specified 1000 Created a chunk of size 1056, which is longer than the specified 1000 Created a chunk of size 1195, which is longer than the specified 1000 Created a chunk of size 3608, which is longer than the specified 1000 Created a chunk of size 1058, which is longer than the specified 1000 Created a chunk of size 1075, which is longer than the specified 1000 Created a chunk of size 1217, which is longer than the specified 1000 Created a chunk of size 1109, which is longer than the specified 1000 Created a chunk of size 1440, which is longer than the specified 1000 Created a chunk of size 1046, which is longer than the specified 1000 Created a chunk of size 1220, which is longer than the specified 1000
https://python.langchain.com/en/latest/use_cases/code/code-analysis-deeplake.html
02980c34832a-6
Created a chunk of size 1220, which is longer than the specified 1000 Created a chunk of size 1403, which is longer than the specified 1000 Created a chunk of size 1241, which is longer than the specified 1000 Created a chunk of size 1427, which is longer than the specified 1000 Created a chunk of size 1049, which is longer than the specified 1000 Created a chunk of size 1580, which is longer than the specified 1000 Created a chunk of size 1565, which is longer than the specified 1000 Created a chunk of size 1131, which is longer than the specified 1000 Created a chunk of size 1425, which is longer than the specified 1000 Created a chunk of size 1054, which is longer than the specified 1000 Created a chunk of size 1027, which is longer than the specified 1000 Created a chunk of size 2559, which is longer than the specified 1000 Created a chunk of size 1028, which is longer than the specified 1000 Created a chunk of size 1382, which is longer than the specified 1000 Created a chunk of size 1888, which is longer than the specified 1000 Created a chunk of size 1475, which is longer than the specified 1000 Created a chunk of size 1652, which is longer than the specified 1000 Created a chunk of size 1891, which is longer than the specified 1000 Created a chunk of size 1899, which is longer than the specified 1000 Created a chunk of size 1021, which is longer than the specified 1000 Created a chunk of size 1085, which is longer than the specified 1000
https://python.langchain.com/en/latest/use_cases/code/code-analysis-deeplake.html
02980c34832a-7
Created a chunk of size 1085, which is longer than the specified 1000 Created a chunk of size 1854, which is longer than the specified 1000 Created a chunk of size 1672, which is longer than the specified 1000 Created a chunk of size 2537, which is longer than the specified 1000 Created a chunk of size 1251, which is longer than the specified 1000 Created a chunk of size 1734, which is longer than the specified 1000 Created a chunk of size 1642, which is longer than the specified 1000 Created a chunk of size 1376, which is longer than the specified 1000 Created a chunk of size 1253, which is longer than the specified 1000 Created a chunk of size 1642, which is longer than the specified 1000 Created a chunk of size 1419, which is longer than the specified 1000 Created a chunk of size 1438, which is longer than the specified 1000 Created a chunk of size 1427, which is longer than the specified 1000 Created a chunk of size 1684, which is longer than the specified 1000 Created a chunk of size 1760, which is longer than the specified 1000 Created a chunk of size 1157, which is longer than the specified 1000 Created a chunk of size 2504, which is longer than the specified 1000 Created a chunk of size 1082, which is longer than the specified 1000 Created a chunk of size 2268, which is longer than the specified 1000 Created a chunk of size 1784, which is longer than the specified 1000 Created a chunk of size 1311, which is longer than the specified 1000
https://python.langchain.com/en/latest/use_cases/code/code-analysis-deeplake.html
02980c34832a-8
Created a chunk of size 1311, which is longer than the specified 1000 Created a chunk of size 2972, which is longer than the specified 1000 Created a chunk of size 1144, which is longer than the specified 1000 Created a chunk of size 1825, which is longer than the specified 1000 Created a chunk of size 1508, which is longer than the specified 1000 Created a chunk of size 2901, which is longer than the specified 1000 Created a chunk of size 1715, which is longer than the specified 1000 Created a chunk of size 1062, which is longer than the specified 1000 Created a chunk of size 1206, which is longer than the specified 1000 Created a chunk of size 1102, which is longer than the specified 1000 Created a chunk of size 1184, which is longer than the specified 1000 Created a chunk of size 1002, which is longer than the specified 1000 Created a chunk of size 1065, which is longer than the specified 1000 Created a chunk of size 1871, which is longer than the specified 1000 Created a chunk of size 1754, which is longer than the specified 1000 Created a chunk of size 2413, which is longer than the specified 1000 Created a chunk of size 1771, which is longer than the specified 1000 Created a chunk of size 2054, which is longer than the specified 1000 Created a chunk of size 2000, which is longer than the specified 1000 Created a chunk of size 2061, which is longer than the specified 1000 Created a chunk of size 1066, which is longer than the specified 1000
https://python.langchain.com/en/latest/use_cases/code/code-analysis-deeplake.html
02980c34832a-9
Created a chunk of size 1066, which is longer than the specified 1000 Created a chunk of size 1419, which is longer than the specified 1000 Created a chunk of size 1368, which is longer than the specified 1000 Created a chunk of size 1008, which is longer than the specified 1000 Created a chunk of size 1227, which is longer than the specified 1000 Created a chunk of size 1745, which is longer than the specified 1000 Created a chunk of size 2296, which is longer than the specified 1000 Created a chunk of size 1083, which is longer than the specified 1000 3477 Then embed chunks and upload them to the DeepLake. This can take several minutes. from langchain.embeddings.openai import OpenAIEmbeddings embeddings = OpenAIEmbeddings() embeddings OpenAIEmbeddings(client=<class 'openai.api_resources.embedding.Embedding'>, model='text-embedding-ada-002', document_model_name='text-embedding-ada-002', query_model_name='text-embedding-ada-002', embedding_ctx_length=8191, openai_api_key=None, openai_organization=None, allowed_special=set(), disallowed_special='all', chunk_size=1000, max_retries=6) from langchain.vectorstores import DeepLake db = DeepLake.from_documents(texts, embeddings, dataset_path=f"hub://{DEEPLAKE_ACCOUNT_NAME}/langchain-code") db Question Answering# First load the dataset, construct the retriever, then construct the Conversational Chain db = DeepLake(dataset_path=f"hub://{DEEPLAKE_ACCOUNT_NAME}/langchain-code", read_only=True, embedding_function=embeddings) -
https://python.langchain.com/en/latest/use_cases/code/code-analysis-deeplake.html
02980c34832a-10
- This dataset can be visualized in Jupyter Notebook by ds.visualize() or at https://app.activeloop.ai/user_name/langchain-code / hub://user_name/langchain-code loaded successfully. Deep Lake Dataset in hub://user_name/langchain-code already exists, loading from the storage Dataset(path='hub://user_name/langchain-code', read_only=True, tensors=['embedding', 'ids', 'metadata', 'text']) tensor htype shape dtype compression ------- ------- ------- ------- ------- embedding generic (3477, 1536) float32 None ids text (3477, 1) str None metadata json (3477, 1) str None text text (3477, 1) str None retriever = db.as_retriever() retriever.search_kwargs['distance_metric'] = 'cos' retriever.search_kwargs['fetch_k'] = 20 retriever.search_kwargs['maximal_marginal_relevance'] = True retriever.search_kwargs['k'] = 20 You can also specify user defined functions using Deep Lake filters def filter(x): # filter based on source code if 'something' in x['text'].data()['value']: return False # filter based on path e.g. extension metadata = x['metadata'].data()['value'] return 'only_this' in metadata['source'] or 'also_that' in metadata['source'] ### turn on below for custom filtering # retriever.search_kwargs['filter'] = filter from langchain.chat_models import ChatOpenAI from langchain.chains import ConversationalRetrievalChain
https://python.langchain.com/en/latest/use_cases/code/code-analysis-deeplake.html
02980c34832a-11
from langchain.chains import ConversationalRetrievalChain model = ChatOpenAI(model='gpt-3.5-turbo') # 'ada' 'gpt-3.5-turbo' 'gpt-4', qa = ConversationalRetrievalChain.from_llm(model,retriever=retriever) questions = [ "What is the class hierarchy?", # "What classes are derived from the Chain class?", # "What classes and functions in the ./langchain/utilities/ forlder are not covered by unit tests?", # "What one improvement do you propose in code in relation to the class herarchy for the Chain class?", ] chat_history = [] for question in questions: result = qa({"question": question, "chat_history": chat_history}) chat_history.append((question, result['answer'])) print(f"-> **Question**: {question} \n") print(f"**Answer**: {result['answer']} \n") -> Question: What is the class hierarchy? Answer: There are several class hierarchies in the provided code, so I’ll list a few: BaseModel -> ConstitutionalPrinciple: ConstitutionalPrinciple is a subclass of BaseModel. BasePromptTemplate -> StringPromptTemplate, AIMessagePromptTemplate, BaseChatPromptTemplate, ChatMessagePromptTemplate, ChatPromptTemplate, HumanMessagePromptTemplate, MessagesPlaceholder, SystemMessagePromptTemplate, FewShotPromptTemplate, FewShotPromptWithTemplates, Prompt, PromptTemplate: All of these classes are subclasses of BasePromptTemplate.
https://python.langchain.com/en/latest/use_cases/code/code-analysis-deeplake.html
02980c34832a-12
APIChain, Chain, MapReduceDocumentsChain, MapRerankDocumentsChain, RefineDocumentsChain, StuffDocumentsChain, HypotheticalDocumentEmbedder, LLMChain, LLMBashChain, LLMCheckerChain, LLMMathChain, LLMRequestsChain, PALChain, QAWithSourcesChain, VectorDBQAWithSourcesChain, VectorDBQA, SQLDatabaseChain: All of these classes are subclasses of Chain. BaseLoader: BaseLoader is a subclass of ABC. BaseTracer -> ChainRun, LLMRun, SharedTracer, ToolRun, Tracer, TracerException, TracerSession: All of these classes are subclasses of BaseTracer. OpenAIEmbeddings, HuggingFaceEmbeddings, CohereEmbeddings, JinaEmbeddings, LlamaCppEmbeddings, HuggingFaceHubEmbeddings, TensorflowHubEmbeddings, SagemakerEndpointEmbeddings, HuggingFaceInstructEmbeddings, SelfHostedEmbeddings, SelfHostedHuggingFaceEmbeddings, SelfHostedHuggingFaceInstructEmbeddings, FakeEmbeddings, AlephAlphaAsymmetricSemanticEmbedding, AlephAlphaSymmetricSemanticEmbedding: All of these classes are subclasses of BaseLLM. -> Question: What classes are derived from the Chain class? Answer: There are multiple classes that are derived from the Chain class. Some of them are: APIChain AnalyzeDocumentChain ChatVectorDBChain CombineDocumentsChain ConstitutionalChain ConversationChain GraphQAChain HypotheticalDocumentEmbedder LLMChain LLMCheckerChain LLMRequestsChain LLMSummarizationCheckerChain MapReduceChain OpenAPIEndpointChain PALChain QAWithSourcesChain RetrievalQA RetrievalQAWithSourcesChain SequentialChain SQLDatabaseChain TransformChain VectorDBQA VectorDBQAWithSourcesChain
https://python.langchain.com/en/latest/use_cases/code/code-analysis-deeplake.html
02980c34832a-13
SequentialChain SQLDatabaseChain TransformChain VectorDBQA VectorDBQAWithSourcesChain There might be more classes that are derived from the Chain class as it is possible to create custom classes that extend the Chain class. -> Question: What classes and functions in the ./langchain/utilities/ forlder are not covered by unit tests? Answer: All classes and functions in the ./langchain/utilities/ folder seem to have unit tests written for them. Contents Design Implementation Integration preparations Prepare data Question Answering By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 25, 2023.
https://python.langchain.com/en/latest/use_cases/code/code-analysis-deeplake.html
9f11795b7a23-0
.ipynb .pdf Wikibase Agent Contents Wikibase Agent Preliminaries API keys and other secrats OpenAI API Key Wikidata user-agent header Enable tracing if desired Tools Item and Property lookup Sparql runner Agent Wrap the tools Prompts Output parser Specify the LLM model Agent and agent executor Run it! Wikibase Agent# This notebook demonstrates a very simple wikibase agent that uses sparql generation. Although this code is intended to work against any wikibase instance, we use http://wikidata.org for testing. If you are interested in wikibases and sparql, please consider helping to improve this agent. Look here for more details and open questions. Preliminaries# API keys and other secrats# We use an .ini file, like this: [OPENAI] OPENAI_API_KEY=xyzzy [WIKIDATA] WIKIDATA_USER_AGENT_HEADER=argle-bargle import configparser config = configparser.ConfigParser() config.read('./secrets.ini') ['./secrets.ini'] OpenAI API Key# An OpenAI API key is required unless you modify the code below to use another LLM provider. openai_api_key = config['OPENAI']['OPENAI_API_KEY'] import os os.environ.update({'OPENAI_API_KEY': openai_api_key}) Wikidata user-agent header# Wikidata policy requires a user-agent header. See https://meta.wikimedia.org/wiki/User-Agent_policy. However, at present this policy is not strictly enforced. wikidata_user_agent_header = None if not config.has_section('WIKIDATA') else config['WIKIDATA']['WIKIDAtA_USER_AGENT_HEADER'] Enable tracing if desired# #import os
https://python.langchain.com/en/latest/use_cases/agents/wikibase_agent.html
9f11795b7a23-1
Enable tracing if desired# #import os #os.environ["LANGCHAIN_HANDLER"] = "langchain" #os.environ["LANGCHAIN_SESSION"] = "default" # Make sure this session actually exists. Tools# Three tools are provided for this simple agent: ItemLookup: for finding the q-number of an item PropertyLookup: for finding the p-number of a property SparqlQueryRunner: for running a sparql query Item and Property lookup# Item and Property lookup are implemented in a single method, using an elastic search endpoint. Not all wikibase instances have it, but wikidata does, and that’s where we’ll start. def get_nested_value(o: dict, path: list) -> any: current = o for key in path: try: current = current[key] except: return None return current import requests from typing import Optional def vocab_lookup(search: str, entity_type: str = "item", url: str = "https://www.wikidata.org/w/api.php", user_agent_header: str = wikidata_user_agent_header, srqiprofile: str = None, ) -> Optional[str]: headers = { 'Accept': 'application/json' } if wikidata_user_agent_header is not None: headers['User-Agent'] = wikidata_user_agent_header if entity_type == "item": srnamespace = 0 srqiprofile = "classic_noboostlinks" if srqiprofile is None else srqiprofile elif entity_type == "property": srnamespace = 120 srqiprofile = "classic" if srqiprofile is None else srqiprofile else:
https://python.langchain.com/en/latest/use_cases/agents/wikibase_agent.html
9f11795b7a23-2
else: raise ValueError("entity_type must be either 'property' or 'item'") params = { "action": "query", "list": "search", "srsearch": search, "srnamespace": srnamespace, "srlimit": 1, "srqiprofile": srqiprofile, "srwhat": 'text', "format": "json" } response = requests.get(url, headers=headers, params=params) if response.status_code == 200: title = get_nested_value(response.json(), ['query', 'search', 0, 'title']) if title is None: return f"I couldn't find any {entity_type} for '{search}'. Please rephrase your request and try again" # if there is a prefix, strip it off return title.split(':')[-1] else: return "Sorry, I got an error. Please try again." print(vocab_lookup("Malin 1")) Q4180017 print(vocab_lookup("instance of", entity_type="property")) P31 print(vocab_lookup("Ceci n'est pas un q-item")) I couldn't find any item for 'Ceci n'est pas un q-item'. Please rephrase your request and try again Sparql runner# This tool runs sparql - by default, wikidata is used. import requests from typing import List, Dict, Any import json def run_sparql(query: str, url='https://query.wikidata.org/sparql', user_agent_header: str = wikidata_user_agent_header) -> List[Dict[str, Any]]: headers = { 'Accept': 'application/json'
https://python.langchain.com/en/latest/use_cases/agents/wikibase_agent.html
9f11795b7a23-3
headers = { 'Accept': 'application/json' } if wikidata_user_agent_header is not None: headers['User-Agent'] = wikidata_user_agent_header response = requests.get(url, headers=headers, params={'query': query, 'format': 'json'}) if response.status_code != 200: return "That query failed. Perhaps you could try a different one?" results = get_nested_value(response.json(),['results', 'bindings']) return json.dumps(results) run_sparql("SELECT (COUNT(?children) as ?count) WHERE { wd:Q1339 wdt:P40 ?children . }") '[{"count": {"datatype": "http://www.w3.org/2001/XMLSchema#integer", "type": "literal", "value": "20"}}]' Agent# Wrap the tools# from langchain.agents import Tool, AgentExecutor, LLMSingleActionAgent, AgentOutputParser from langchain.prompts import StringPromptTemplate from langchain import OpenAI, LLMChain from typing import List, Union from langchain.schema import AgentAction, AgentFinish import re # Define which tools the agent can use to answer user queries tools = [ Tool( name = "ItemLookup", func=(lambda x: vocab_lookup(x, entity_type="item")), description="useful for when you need to know the q-number for an item" ), Tool( name = "PropertyLookup", func=(lambda x: vocab_lookup(x, entity_type="property")), description="useful for when you need to know the p-number for a property" ), Tool( name = "SparqlQueryRunner", func=run_sparql,
https://python.langchain.com/en/latest/use_cases/agents/wikibase_agent.html
9f11795b7a23-4
name = "SparqlQueryRunner", func=run_sparql, description="useful for getting results from a wikibase" ) ] Prompts# # Set up the base template template = """ Answer the following questions by running a sparql query against a wikibase where the p and q items are completely unknown to you. You will need to discover the p and q items before you can generate the sparql. Do not assume you know the p and q items for any concepts. Always use tools to find all p and q items. After you generate the sparql, you should run it. The results will be returned in json. Summarize the json results in natural language. You may assume the following prefixes: PREFIX wd: <http://www.wikidata.org/entity/> PREFIX wdt: <http://www.wikidata.org/prop/direct/> PREFIX p: <http://www.wikidata.org/prop/> PREFIX ps: <http://www.wikidata.org/prop/statement/> When generating sparql: * Try to avoid "count" and "filter" queries if possible * Never enclose the sparql in back-quotes You have access to the following tools: {tools} Use the following format: Question: the input question for which you must provide a natural language answer Thought: you should always think about what to do Action: the action to take, should be one of [{tool_names}] Action Input: the input to the action Observation: the result of the action ... (this Thought/Action/Action Input/Observation can repeat N times) Thought: I now know the final answer Final Answer: the final answer to the original input question Question: {input} {agent_scratchpad}"""
https://python.langchain.com/en/latest/use_cases/agents/wikibase_agent.html
9f11795b7a23-5
Question: {input} {agent_scratchpad}""" # Set up a prompt template class CustomPromptTemplate(StringPromptTemplate): # The template to use template: str # The list of tools available tools: List[Tool] def format(self, **kwargs) -> str: # Get the intermediate steps (AgentAction, Observation tuples) # Format them in a particular way intermediate_steps = kwargs.pop("intermediate_steps") thoughts = "" for action, observation in intermediate_steps: thoughts += action.log thoughts += f"\nObservation: {observation}\nThought: " # Set the agent_scratchpad variable to that value kwargs["agent_scratchpad"] = thoughts # Create a tools variable from the list of tools provided kwargs["tools"] = "\n".join([f"{tool.name}: {tool.description}" for tool in self.tools]) # Create a list of tool names for the tools provided kwargs["tool_names"] = ", ".join([tool.name for tool in self.tools]) return self.template.format(**kwargs) prompt = CustomPromptTemplate( template=template, tools=tools, # This omits the `agent_scratchpad`, `tools`, and `tool_names` variables because those are generated dynamically # This includes the `intermediate_steps` variable because that is needed input_variables=["input", "intermediate_steps"] ) Output parser# This is unchanged from langchain docs class CustomOutputParser(AgentOutputParser): def parse(self, llm_output: str) -> Union[AgentAction, AgentFinish]: # Check if agent should finish if "Final Answer:" in llm_output: return AgentFinish(
https://python.langchain.com/en/latest/use_cases/agents/wikibase_agent.html
9f11795b7a23-6
if "Final Answer:" in llm_output: return AgentFinish( # Return values is generally always a dictionary with a single `output` key # It is not recommended to try anything else at the moment :) return_values={"output": llm_output.split("Final Answer:")[-1].strip()}, log=llm_output, ) # Parse out the action and action input regex = r"Action: (.*?)[\n]*Action Input:[\s]*(.*)" match = re.search(regex, llm_output, re.DOTALL) if not match: raise ValueError(f"Could not parse LLM output: `{llm_output}`") action = match.group(1).strip() action_input = match.group(2) # Return the action and action input return AgentAction(tool=action, tool_input=action_input.strip(" ").strip('"'), log=llm_output) output_parser = CustomOutputParser() Specify the LLM model# from langchain.chat_models import ChatOpenAI llm = ChatOpenAI(model="gpt-4", temperature=0) Agent and agent executor# # LLM chain consisting of the LLM and a prompt llm_chain = LLMChain(llm=llm, prompt=prompt) tool_names = [tool.name for tool in tools] agent = LLMSingleActionAgent( llm_chain=llm_chain, output_parser=output_parser, stop=["\nObservation:"], allowed_tools=tool_names ) agent_executor = AgentExecutor.from_agent_and_tools(agent=agent, tools=tools, verbose=True) Run it!# # If you prefer in-line tracing, uncomment this line # agent_executor.agent.llm_chain.verbose = True
https://python.langchain.com/en/latest/use_cases/agents/wikibase_agent.html
9f11795b7a23-7
# agent_executor.agent.llm_chain.verbose = True agent_executor.run("How many children did J.S. Bach have?") > Entering new AgentExecutor chain... Thought: I need to find the Q number for J.S. Bach. Action: ItemLookup Action Input: J.S. Bach Observation:Q1339I need to find the P number for children. Action: PropertyLookup Action Input: children Observation:P1971Now I can query the number of children J.S. Bach had. Action: SparqlQueryRunner Action Input: SELECT ?children WHERE { wd:Q1339 wdt:P1971 ?children } Observation:[{"children": {"datatype": "http://www.w3.org/2001/XMLSchema#decimal", "type": "literal", "value": "20"}}]I now know the final answer. Final Answer: J.S. Bach had 20 children. > Finished chain. 'J.S. Bach had 20 children.' agent_executor.run("What is the Basketball-Reference.com NBA player ID of Hakeem Olajuwon?") > Entering new AgentExecutor chain... Thought: To find Hakeem Olajuwon's Basketball-Reference.com NBA player ID, I need to first find his Wikidata item (Q-number) and then query for the relevant property (P-number). Action: ItemLookup Action Input: Hakeem Olajuwon Observation:Q273256Now that I have Hakeem Olajuwon's Wikidata item (Q273256), I need to find the P-number for the Basketball-Reference.com NBA player ID property. Action: PropertyLookup Action Input: Basketball-Reference.com NBA player ID
https://python.langchain.com/en/latest/use_cases/agents/wikibase_agent.html
9f11795b7a23-8
Action: PropertyLookup Action Input: Basketball-Reference.com NBA player ID Observation:P2685Now that I have both the Q-number for Hakeem Olajuwon (Q273256) and the P-number for the Basketball-Reference.com NBA player ID property (P2685), I can run a SPARQL query to get the ID value. Action: SparqlQueryRunner Action Input: SELECT ?playerID WHERE { wd:Q273256 wdt:P2685 ?playerID . } Observation:[{"playerID": {"type": "literal", "value": "o/olajuha01"}}]I now know the final answer Final Answer: Hakeem Olajuwon's Basketball-Reference.com NBA player ID is "o/olajuha01". > Finished chain. 'Hakeem Olajuwon\'s Basketball-Reference.com NBA player ID is "o/olajuha01".' Contents Wikibase Agent Preliminaries API keys and other secrats OpenAI API Key Wikidata user-agent header Enable tracing if desired Tools Item and Property lookup Sparql runner Agent Wrap the tools Prompts Output parser Specify the LLM model Agent and agent executor Run it! By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 25, 2023.
https://python.langchain.com/en/latest/use_cases/agents/wikibase_agent.html
a796a62ee6c5-0
.ipynb .pdf Custom Agent with PlugIn Retrieval Contents Set up environment Setup LLM Set up plugins Tool Retriever Prompt Template Output Parser Set up LLM, stop sequence, and the agent Use the Agent Custom Agent with PlugIn Retrieval# This notebook combines two concepts in order to build a custom agent that can interact with AI Plugins: Custom Agent with Retrieval: This introduces the concept of retrieving many tools, which is useful when trying to work with arbitrarily many plugins. Natural Language API Chains: This creates Natural Language wrappers around OpenAPI endpoints. This is useful because (1) plugins use OpenAPI endpoints under the hood, (2) wrapping them in an NLAChain allows the router agent to call it more easily. The novel idea introduced in this notebook is the idea of using retrieval to select not the tools explicitly, but the set of OpenAPI specs to use. We can then generate tools from those OpenAPI specs. The use case for this is when trying to get agents to use plugins. It may be more efficient to choose plugins first, then the endpoints, rather than the endpoints directly. This is because the plugins may contain more useful information for selection. Set up environment# Do necessary imports, etc. from langchain.agents import Tool, AgentExecutor, LLMSingleActionAgent, AgentOutputParser from langchain.prompts import StringPromptTemplate from langchain import OpenAI, SerpAPIWrapper, LLMChain from typing import List, Union from langchain.schema import AgentAction, AgentFinish from langchain.agents.agent_toolkits import NLAToolkit from langchain.tools.plugin import AIPlugin import re Setup LLM# llm = OpenAI(temperature=0) Set up plugins# Load and index plugins urls = [
https://python.langchain.com/en/latest/use_cases/agents/custom_agent_with_plugin_retrieval.html
a796a62ee6c5-1
Set up plugins# Load and index plugins urls = [ "https://datasette.io/.well-known/ai-plugin.json", "https://api.speak.com/.well-known/ai-plugin.json", "https://www.wolframalpha.com/.well-known/ai-plugin.json", "https://www.zapier.com/.well-known/ai-plugin.json", "https://www.klarna.com/.well-known/ai-plugin.json", "https://www.joinmilo.com/.well-known/ai-plugin.json", "https://slack.com/.well-known/ai-plugin.json", "https://schooldigger.com/.well-known/ai-plugin.json", ] AI_PLUGINS = [AIPlugin.from_url(url) for url in urls] Tool Retriever# We will use a vectorstore to create embeddings for each tool description. Then, for an incoming query we can create embeddings for that query and do a similarity search for relevant tools. from langchain.vectorstores import FAISS from langchain.embeddings import OpenAIEmbeddings from langchain.schema import Document embeddings = OpenAIEmbeddings() docs = [ Document(page_content=plugin.description_for_model, metadata={"plugin_name": plugin.name_for_model} ) for plugin in AI_PLUGINS ] vector_store = FAISS.from_documents(docs, embeddings) toolkits_dict = {plugin.name_for_model: NLAToolkit.from_llm_and_ai_plugin(llm, plugin) for plugin in AI_PLUGINS} Attempting to load an OpenAPI 3.0.1 spec. This may result in degraded performance. Convert your OpenAPI spec to 3.1.* spec for better support.
https://python.langchain.com/en/latest/use_cases/agents/custom_agent_with_plugin_retrieval.html
a796a62ee6c5-2
Attempting to load an OpenAPI 3.0.1 spec. This may result in degraded performance. Convert your OpenAPI spec to 3.1.* spec for better support. Attempting to load an OpenAPI 3.0.1 spec. This may result in degraded performance. Convert your OpenAPI spec to 3.1.* spec for better support. Attempting to load an OpenAPI 3.0.2 spec. This may result in degraded performance. Convert your OpenAPI spec to 3.1.* spec for better support. Attempting to load an OpenAPI 3.0.1 spec. This may result in degraded performance. Convert your OpenAPI spec to 3.1.* spec for better support. Attempting to load an OpenAPI 3.0.1 spec. This may result in degraded performance. Convert your OpenAPI spec to 3.1.* spec for better support. Attempting to load an OpenAPI 3.0.1 spec. This may result in degraded performance. Convert your OpenAPI spec to 3.1.* spec for better support. Attempting to load an OpenAPI 3.0.1 spec. This may result in degraded performance. Convert your OpenAPI spec to 3.1.* spec for better support. Attempting to load a Swagger 2.0 spec. This may result in degraded performance. Convert your OpenAPI spec to 3.1.* spec for better support. retriever = vector_store.as_retriever() def get_tools(query): # Get documents, which contain the Plugins to use docs = retriever.get_relevant_documents(query) # Get the toolkits, one for each plugin tool_kits = [toolkits_dict[d.metadata["plugin_name"]] for d in docs] # Get the tools: a separate NLAChain for each endpoint
https://python.langchain.com/en/latest/use_cases/agents/custom_agent_with_plugin_retrieval.html
a796a62ee6c5-3
# Get the tools: a separate NLAChain for each endpoint tools = [] for tk in tool_kits: tools.extend(tk.nla_tools) return tools We can now test this retriever to see if it seems to work. tools = get_tools("What could I do today with my kiddo") [t.name for t in tools] ['Milo.askMilo', 'Zapier_Natural_Language_Actions_(NLA)_API_(Dynamic)_-_Beta.search_all_actions', 'Zapier_Natural_Language_Actions_(NLA)_API_(Dynamic)_-_Beta.preview_a_zap', 'Zapier_Natural_Language_Actions_(NLA)_API_(Dynamic)_-_Beta.get_configuration_link', 'Zapier_Natural_Language_Actions_(NLA)_API_(Dynamic)_-_Beta.list_exposed_actions', 'SchoolDigger_API_V2.0.Autocomplete_GetSchools', 'SchoolDigger_API_V2.0.Districts_GetAllDistricts2', 'SchoolDigger_API_V2.0.Districts_GetDistrict2', 'SchoolDigger_API_V2.0.Rankings_GetSchoolRank2', 'SchoolDigger_API_V2.0.Rankings_GetRank_District', 'SchoolDigger_API_V2.0.Schools_GetAllSchools20', 'SchoolDigger_API_V2.0.Schools_GetSchool20', 'Speak.translate', 'Speak.explainPhrase', 'Speak.explainTask'] tools = get_tools("what shirts can i buy?") [t.name for t in tools] ['Open_AI_Klarna_product_Api.productsUsingGET', 'Milo.askMilo',
https://python.langchain.com/en/latest/use_cases/agents/custom_agent_with_plugin_retrieval.html
a796a62ee6c5-4
['Open_AI_Klarna_product_Api.productsUsingGET', 'Milo.askMilo', 'Zapier_Natural_Language_Actions_(NLA)_API_(Dynamic)_-_Beta.search_all_actions', 'Zapier_Natural_Language_Actions_(NLA)_API_(Dynamic)_-_Beta.preview_a_zap', 'Zapier_Natural_Language_Actions_(NLA)_API_(Dynamic)_-_Beta.get_configuration_link', 'Zapier_Natural_Language_Actions_(NLA)_API_(Dynamic)_-_Beta.list_exposed_actions', 'SchoolDigger_API_V2.0.Autocomplete_GetSchools', 'SchoolDigger_API_V2.0.Districts_GetAllDistricts2', 'SchoolDigger_API_V2.0.Districts_GetDistrict2', 'SchoolDigger_API_V2.0.Rankings_GetSchoolRank2', 'SchoolDigger_API_V2.0.Rankings_GetRank_District', 'SchoolDigger_API_V2.0.Schools_GetAllSchools20', 'SchoolDigger_API_V2.0.Schools_GetSchool20'] Prompt Template# The prompt template is pretty standard, because we’re not actually changing that much logic in the actual prompt template, but rather we are just changing how retrieval is done. # Set up the base template template = """Answer the following questions as best you can, but speaking as a pirate might speak. You have access to the following tools: {tools} Use the following format: Question: the input question you must answer Thought: you should always think about what to do Action: the action to take, should be one of [{tool_names}] Action Input: the input to the action Observation: the result of the action
https://python.langchain.com/en/latest/use_cases/agents/custom_agent_with_plugin_retrieval.html
a796a62ee6c5-5
Action Input: the input to the action Observation: the result of the action ... (this Thought/Action/Action Input/Observation can repeat N times) Thought: I now know the final answer Final Answer: the final answer to the original input question Begin! Remember to speak as a pirate when giving your final answer. Use lots of "Arg"s Question: {input} {agent_scratchpad}""" The custom prompt template now has the concept of a tools_getter, which we call on the input to select the tools to use from typing import Callable # Set up a prompt template class CustomPromptTemplate(StringPromptTemplate): # The template to use template: str ############## NEW ###################### # The list of tools available tools_getter: Callable def format(self, **kwargs) -> str: # Get the intermediate steps (AgentAction, Observation tuples) # Format them in a particular way intermediate_steps = kwargs.pop("intermediate_steps") thoughts = "" for action, observation in intermediate_steps: thoughts += action.log thoughts += f"\nObservation: {observation}\nThought: " # Set the agent_scratchpad variable to that value kwargs["agent_scratchpad"] = thoughts ############## NEW ###################### tools = self.tools_getter(kwargs["input"]) # Create a tools variable from the list of tools provided kwargs["tools"] = "\n".join([f"{tool.name}: {tool.description}" for tool in tools]) # Create a list of tool names for the tools provided kwargs["tool_names"] = ", ".join([tool.name for tool in tools]) return self.template.format(**kwargs) prompt = CustomPromptTemplate( template=template,
https://python.langchain.com/en/latest/use_cases/agents/custom_agent_with_plugin_retrieval.html
a796a62ee6c5-6
prompt = CustomPromptTemplate( template=template, tools_getter=get_tools, # This omits the `agent_scratchpad`, `tools`, and `tool_names` variables because those are generated dynamically # This includes the `intermediate_steps` variable because that is needed input_variables=["input", "intermediate_steps"] ) Output Parser# The output parser is unchanged from the previous notebook, since we are not changing anything about the output format. class CustomOutputParser(AgentOutputParser): def parse(self, llm_output: str) -> Union[AgentAction, AgentFinish]: # Check if agent should finish if "Final Answer:" in llm_output: return AgentFinish( # Return values is generally always a dictionary with a single `output` key # It is not recommended to try anything else at the moment :) return_values={"output": llm_output.split("Final Answer:")[-1].strip()}, log=llm_output, ) # Parse out the action and action input regex = r"Action\s*\d*\s*:(.*?)\nAction\s*\d*\s*Input\s*\d*\s*:[\s]*(.*)" match = re.search(regex, llm_output, re.DOTALL) if not match: raise ValueError(f"Could not parse LLM output: `{llm_output}`") action = match.group(1).strip() action_input = match.group(2) # Return the action and action input return AgentAction(tool=action, tool_input=action_input.strip(" ").strip('"'), log=llm_output) output_parser = CustomOutputParser() Set up LLM, stop sequence, and the agent# Also the same as the previous notebook
https://python.langchain.com/en/latest/use_cases/agents/custom_agent_with_plugin_retrieval.html
a796a62ee6c5-7
Set up LLM, stop sequence, and the agent# Also the same as the previous notebook llm = OpenAI(temperature=0) # LLM chain consisting of the LLM and a prompt llm_chain = LLMChain(llm=llm, prompt=prompt) tool_names = [tool.name for tool in tools] agent = LLMSingleActionAgent( llm_chain=llm_chain, output_parser=output_parser, stop=["\nObservation:"], allowed_tools=tool_names ) Use the Agent# Now we can use it! agent_executor = AgentExecutor.from_agent_and_tools(agent=agent, tools=tools, verbose=True) agent_executor.run("what shirts can i buy?") > Entering new AgentExecutor chain... Thought: I need to find a product API Action: Open_AI_Klarna_product_Api.productsUsingGET Action Input: shirts Observation:I found 10 shirts from the API response. They range in price from $9.99 to $450.00 and come in a variety of materials, colors, and patterns. I now know what shirts I can buy Final Answer: Arg, I found 10 shirts from the API response. They range in price from $9.99 to $450.00 and come in a variety of materials, colors, and patterns. > Finished chain. 'Arg, I found 10 shirts from the API response. They range in price from $9.99 to $450.00 and come in a variety of materials, colors, and patterns.' Contents Set up environment Setup LLM Set up plugins Tool Retriever Prompt Template Output Parser Set up LLM, stop sequence, and the agent Use the Agent By Harrison Chase © Copyright 2023, Harrison Chase.
https://python.langchain.com/en/latest/use_cases/agents/custom_agent_with_plugin_retrieval.html
a796a62ee6c5-8
Use the Agent By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 25, 2023.
https://python.langchain.com/en/latest/use_cases/agents/custom_agent_with_plugin_retrieval.html
a256932054ac-0
.ipynb .pdf Benchmarking Template Contents Loading the data Setting up a chain Make a prediction Make many predictions Evaluate performance Benchmarking Template# This is an example notebook that can be used to create a benchmarking notebook for a task of your choice. Evaluation is really hard, and so we greatly welcome any contributions that can make it easier for people to experiment It is highly reccomended that you do any evaluation/benchmarking with tracing enabled. See here for an explanation of what tracing is and how to set it up. # Comment this out if you are NOT using tracing import os os.environ["LANGCHAIN_HANDLER"] = "langchain" Loading the data# First, let’s load the data. # This notebook should so how to load the dataset from LangChainDatasets on Hugging Face # Please upload your dataset to https://huggingface.co/LangChainDatasets # The value passed into `load_dataset` should NOT have the `LangChainDatasets/` prefix from langchain.evaluation.loading import load_dataset dataset = load_dataset("TODO") Setting up a chain# This next section should have an example of setting up a chain that can be run on this dataset. Make a prediction# First, we can make predictions one datapoint at a time. Doing it at this level of granularity allows use to explore the outputs in detail, and also is a lot cheaper than running over multiple datapoints # Example of running the chain on a single datapoint (`dataset[0]`) goes here Make many predictions# Now we can make predictions. # Example of running the chain on many predictions goes here # Sometimes its as simple as `chain.apply(dataset)` # Othertimes you may want to write a for loop to catch errors Evaluate performance#
https://python.langchain.com/en/latest/use_cases/evaluation/benchmarking_template.html
a256932054ac-1
# Othertimes you may want to write a for loop to catch errors Evaluate performance# Any guide to evaluating performance in a more systematic manner goes here. previous Agent VectorDB Question Answering Benchmarking next Data Augmented Question Answering Contents Loading the data Setting up a chain Make a prediction Make many predictions Evaluate performance By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 25, 2023.
https://python.langchain.com/en/latest/use_cases/evaluation/benchmarking_template.html
4d6a928947d1-0
.ipynb .pdf Question Answering Contents Setup Examples Predictions Evaluation Customize Prompt Evaluation without Ground Truth Comparing to other evaluation metrics Question Answering# This notebook covers how to evaluate generic question answering problems. This is a situation where you have an example containing a question and its corresponding ground truth answer, and you want to measure how well the language model does at answering those questions. Setup# For demonstration purposes, we will just evaluate a simple question answering system that only evaluates the model’s internal knowledge. Please see other notebooks for examples where it evaluates how the model does at question answering over data not present in what the model was trained on. from langchain.prompts import PromptTemplate from langchain.chains import LLMChain from langchain.llms import OpenAI prompt = PromptTemplate(template="Question: {question}\nAnswer:", input_variables=["question"]) llm = OpenAI(model_name="text-davinci-003", temperature=0) chain = LLMChain(llm=llm, prompt=prompt) Examples# For this purpose, we will just use two simple hardcoded examples, but see other notebooks for tips on how to get and/or generate these examples. examples = [ { "question": "Roger has 5 tennis balls. He buys 2 more cans of tennis balls. Each can has 3 tennis balls. How many tennis balls does he have now?", "answer": "11" }, { "question": 'Is the following sentence plausible? "Joao Moutinho caught the screen pass in the NFC championship."', "answer": "No" } ] Predictions# We can now make and inspect the predictions for these questions. predictions = chain.apply(examples) predictions [{'text': ' 11 tennis balls'},
https://python.langchain.com/en/latest/use_cases/evaluation/question_answering.html
4d6a928947d1-1
predictions = chain.apply(examples) predictions [{'text': ' 11 tennis balls'}, {'text': ' No, this sentence is not plausible. Joao Moutinho is a professional soccer player, not an American football player, so it is not likely that he would be catching a screen pass in the NFC championship.'}] Evaluation# We can see that if we tried to just do exact match on the answer answers (11 and No) they would not match what the language model answered. However, semantically the language model is correct in both cases. In order to account for this, we can use a language model itself to evaluate the answers. from langchain.evaluation.qa import QAEvalChain llm = OpenAI(temperature=0) eval_chain = QAEvalChain.from_llm(llm) graded_outputs = eval_chain.evaluate(examples, predictions, question_key="question", prediction_key="text") for i, eg in enumerate(examples): print(f"Example {i}:") print("Question: " + eg['question']) print("Real Answer: " + eg['answer']) print("Predicted Answer: " + predictions[i]['text']) print("Predicted Grade: " + graded_outputs[i]['text']) print() Example 0: Question: Roger has 5 tennis balls. He buys 2 more cans of tennis balls. Each can has 3 tennis balls. How many tennis balls does he have now? Real Answer: 11 Predicted Answer: 11 tennis balls Predicted Grade: CORRECT Example 1: Question: Is the following sentence plausible? "Joao Moutinho caught the screen pass in the NFC championship." Real Answer: No
https://python.langchain.com/en/latest/use_cases/evaluation/question_answering.html
4d6a928947d1-2
Real Answer: No Predicted Answer: No, this sentence is not plausible. Joao Moutinho is a professional soccer player, not an American football player, so it is not likely that he would be catching a screen pass in the NFC championship. Predicted Grade: CORRECT Customize Prompt# You can also customize the prompt that is used. Here is an example prompting it using a score from 0 to 10. The custom prompt requires 3 input variables: “query”, “answer” and “result”. Where “query” is the question, “answer” is the ground truth answer, and “result” is the predicted answer. from langchain.prompts.prompt import PromptTemplate _PROMPT_TEMPLATE = """You are an expert professor specialized in grading students' answers to questions. You are grading the following question: {query} Here is the real answer: {answer} You are grading the following predicted answer: {result} What grade do you give from 0 to 10, where 0 is the lowest (very low similarity) and 10 is the highest (very high similarity)? """ PROMPT = PromptTemplate(input_variables=["query", "answer", "result"], template=_PROMPT_TEMPLATE) evalchain = QAEvalChain.from_llm(llm=llm,prompt=PROMPT) evalchain.evaluate(examples, predictions, question_key="question", answer_key="answer", prediction_key="text") Evaluation without Ground Truth# Its possible to evaluate question answering systems without ground truth. You would need a "context" input that reflects what the information the LLM uses to answer the question. This context can be obtained by any retreival system. Here’s an example of how it works: context_examples = [ { "question": "How old am I?",
https://python.langchain.com/en/latest/use_cases/evaluation/question_answering.html
4d6a928947d1-3
context_examples = [ { "question": "How old am I?", "context": "I am 30 years old. I live in New York and take the train to work everyday.", }, { "question": 'Who won the NFC championship game in 2023?"', "context": "NFC Championship Game 2023: Philadelphia Eagles 31, San Francisco 49ers 7" } ] QA_PROMPT = "Answer the question based on the context\nContext:{context}\nQuestion:{question}\nAnswer:" template = PromptTemplate(input_variables=["context", "question"], template=QA_PROMPT) qa_chain = LLMChain(llm=llm, prompt=template) predictions = qa_chain.apply(context_examples) predictions [{'text': 'You are 30 years old.'}, {'text': ' The Philadelphia Eagles won the NFC championship game in 2023.'}] from langchain.evaluation.qa import ContextQAEvalChain eval_chain = ContextQAEvalChain.from_llm(llm) graded_outputs = eval_chain.evaluate(context_examples, predictions, question_key="question", prediction_key="text") graded_outputs [{'text': ' CORRECT'}, {'text': ' CORRECT'}] Comparing to other evaluation metrics# We can compare the evaluation results we get to other common evaluation metrics. To do this, let’s load some evaluation metrics from HuggingFace’s evaluate package. # Some data munging to get the examples in the right format for i, eg in enumerate(examples): eg['id'] = str(i) eg['answers'] = {"text": [eg['answer']], "answer_start": [0]} predictions[i]['id'] = str(i)
https://python.langchain.com/en/latest/use_cases/evaluation/question_answering.html
4d6a928947d1-4
predictions[i]['id'] = str(i) predictions[i]['prediction_text'] = predictions[i]['text'] for p in predictions: del p['text'] new_examples = examples.copy() for eg in new_examples: del eg ['question'] del eg['answer'] from evaluate import load squad_metric = load("squad") results = squad_metric.compute( references=new_examples, predictions=predictions, ) results {'exact_match': 0.0, 'f1': 28.125} previous QA Generation next SQL Question Answering Benchmarking: Chinook Contents Setup Examples Predictions Evaluation Customize Prompt Evaluation without Ground Truth Comparing to other evaluation metrics By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 25, 2023.
https://python.langchain.com/en/latest/use_cases/evaluation/question_answering.html
a1952c3cd61b-0
.ipynb .pdf Data Augmented Question Answering Contents Setup Examples Evaluate Evaluate with Other Metrics Data Augmented Question Answering# This notebook uses some generic prompts/language models to evaluate an question answering system that uses other sources of data besides what is in the model. For example, this can be used to evaluate a question answering system over your proprietary data. Setup# Let’s set up an example with our favorite example - the state of the union address. from langchain.embeddings.openai import OpenAIEmbeddings from langchain.vectorstores import Chroma from langchain.text_splitter import CharacterTextSplitter from langchain.llms import OpenAI from langchain.chains import RetrievalQA from langchain.document_loaders import TextLoader loader = TextLoader('../../modules/state_of_the_union.txt') documents = loader.load() text_splitter = CharacterTextSplitter(chunk_size=1000, chunk_overlap=0) texts = text_splitter.split_documents(documents) embeddings = OpenAIEmbeddings() docsearch = Chroma.from_documents(texts, embeddings) qa = RetrievalQA.from_llm(llm=OpenAI(), retriever=docsearch.as_retriever()) Running Chroma using direct local API. Using DuckDB in-memory for database. Data will be transient. Examples# Now we need some examples to evaluate. We can do this in two ways: Hard code some examples ourselves Generate examples automatically, using a language model # Hard-coded examples examples = [ { "query": "What did the president say about Ketanji Brown Jackson", "answer": "He praised her legal ability and said he nominated her for the supreme court." }, { "query": "What did the president say about Michael Jackson", "answer": "Nothing"
https://python.langchain.com/en/latest/use_cases/evaluation/data_augmented_question_answering.html
a1952c3cd61b-1
"answer": "Nothing" } ] # Generated examples from langchain.evaluation.qa import QAGenerateChain example_gen_chain = QAGenerateChain.from_llm(OpenAI()) new_examples = example_gen_chain.apply_and_parse([{"doc": t} for t in texts[:5]]) new_examples [{'query': 'According to the document, what did Vladimir Putin miscalculate?', 'answer': 'He miscalculated that he could roll into Ukraine and the world would roll over.'}, {'query': 'Who is the Ukrainian Ambassador to the United States?', 'answer': 'The Ukrainian Ambassador to the United States is here tonight.'}, {'query': 'How many countries were part of the coalition formed to confront Putin?', 'answer': '27 members of the European Union, France, Germany, Italy, the United Kingdom, Canada, Japan, Korea, Australia, New Zealand, and many others, even Switzerland.'}, {'query': 'What action is the U.S. Department of Justice taking to target Russian oligarchs?', 'answer': 'The U.S. Department of Justice is assembling a dedicated task force to go after the crimes of Russian oligarchs and joining with European allies to find and seize their yachts, luxury apartments, and private jets.'}, {'query': 'How much direct assistance is the United States providing to Ukraine?', 'answer': 'The United States is providing more than $1 Billion in direct assistance to Ukraine.'}] # Combine examples examples += new_examples Evaluate# Now that we have examples, we can use the question answering evaluator to evaluate our question answering chain. from langchain.evaluation.qa import QAEvalChain predictions = qa.apply(examples) llm = OpenAI(temperature=0) eval_chain = QAEvalChain.from_llm(llm)
https://python.langchain.com/en/latest/use_cases/evaluation/data_augmented_question_answering.html
a1952c3cd61b-2
eval_chain = QAEvalChain.from_llm(llm) graded_outputs = eval_chain.evaluate(examples, predictions) for i, eg in enumerate(examples): print(f"Example {i}:") print("Question: " + predictions[i]['query']) print("Real Answer: " + predictions[i]['answer']) print("Predicted Answer: " + predictions[i]['result']) print("Predicted Grade: " + graded_outputs[i]['text']) print() Example 0: Question: What did the president say about Ketanji Brown Jackson Real Answer: He praised her legal ability and said he nominated her for the supreme court. Predicted Answer: The president said that she is one of the nation's top legal minds, a former top litigator in private practice, a former federal public defender, and from a family of public school educators and police officers. He also said that she is a consensus builder and that she has received a broad range of support from the Fraternal Order of Police to former judges appointed by both Democrats and Republicans. Predicted Grade: CORRECT Example 1: Question: What did the president say about Michael Jackson Real Answer: Nothing Predicted Answer: The president did not mention Michael Jackson in this speech. Predicted Grade: CORRECT Example 2: Question: According to the document, what did Vladimir Putin miscalculate? Real Answer: He miscalculated that he could roll into Ukraine and the world would roll over. Predicted Answer: Putin miscalculated that the world would roll over when he rolled into Ukraine. Predicted Grade: CORRECT Example 3: Question: Who is the Ukrainian Ambassador to the United States? Real Answer: The Ukrainian Ambassador to the United States is here tonight. Predicted Answer: I don't know.
https://python.langchain.com/en/latest/use_cases/evaluation/data_augmented_question_answering.html
a1952c3cd61b-3
Predicted Answer: I don't know. Predicted Grade: INCORRECT Example 4: Question: How many countries were part of the coalition formed to confront Putin? Real Answer: 27 members of the European Union, France, Germany, Italy, the United Kingdom, Canada, Japan, Korea, Australia, New Zealand, and many others, even Switzerland. Predicted Answer: The coalition included freedom-loving nations from Europe and the Americas to Asia and Africa, 27 members of the European Union including France, Germany, Italy, the United Kingdom, Canada, Japan, Korea, Australia, New Zealand, and many others, even Switzerland. Predicted Grade: INCORRECT Example 5: Question: What action is the U.S. Department of Justice taking to target Russian oligarchs? Real Answer: The U.S. Department of Justice is assembling a dedicated task force to go after the crimes of Russian oligarchs and joining with European allies to find and seize their yachts, luxury apartments, and private jets. Predicted Answer: The U.S. Department of Justice is assembling a dedicated task force to go after the crimes of Russian oligarchs and to find and seize their yachts, luxury apartments, and private jets. Predicted Grade: INCORRECT Example 6: Question: How much direct assistance is the United States providing to Ukraine? Real Answer: The United States is providing more than $1 Billion in direct assistance to Ukraine. Predicted Answer: The United States is providing more than $1 billion in direct assistance to Ukraine. Predicted Grade: CORRECT Evaluate with Other Metrics#
https://python.langchain.com/en/latest/use_cases/evaluation/data_augmented_question_answering.html
a1952c3cd61b-4
Predicted Grade: CORRECT Evaluate with Other Metrics# In addition to predicting whether the answer is correct or incorrect using a language model, we can also use other metrics to get a more nuanced view on the quality of the answers. To do so, we can use the Critique library, which allows for simple calculation of various metrics over generated text. First you can get an API key from the Inspired Cognition Dashboard and do some setup: export INSPIREDCO_API_KEY="..." pip install inspiredco import inspiredco.critique import os critique = inspiredco.critique.Critique(api_key=os.environ['INSPIREDCO_API_KEY']) Then run the following code to set up the configuration and calculate the ROUGE, chrf, BERTScore, and UniEval (you can choose other metrics too): metrics = { "rouge": { "metric": "rouge", "config": {"variety": "rouge_l"}, }, "chrf": { "metric": "chrf", "config": {}, }, "bert_score": { "metric": "bert_score", "config": {"model": "bert-base-uncased"}, }, "uni_eval": { "metric": "uni_eval", "config": {"task": "summarization", "evaluation_aspect": "relevance"}, }, } critique_data = [ {"target": pred['result'], "references": [pred['answer']]} for pred in predictions ] eval_results = { k: critique.evaluate(dataset=critique_data, metric=v["metric"], config=v["config"]) for k, v in metrics.items() }
https://python.langchain.com/en/latest/use_cases/evaluation/data_augmented_question_answering.html
a1952c3cd61b-5
for k, v in metrics.items() } Finally, we can print out the results. We can see that overall the scores are higher when the output is semantically correct, and also when the output closely matches with the gold-standard answer. for i, eg in enumerate(examples): score_string = ", ".join([f"{k}={v['examples'][i]['value']:.4f}" for k, v in eval_results.items()]) print(f"Example {i}:") print("Question: " + predictions[i]['query']) print("Real Answer: " + predictions[i]['answer']) print("Predicted Answer: " + predictions[i]['result']) print("Predicted Scores: " + score_string) print() Example 0: Question: What did the president say about Ketanji Brown Jackson Real Answer: He praised her legal ability and said he nominated her for the supreme court. Predicted Answer: The president said that she is one of the nation's top legal minds, a former top litigator in private practice, a former federal public defender, and from a family of public school educators and police officers. He also said that she is a consensus builder and that she has received a broad range of support from the Fraternal Order of Police to former judges appointed by both Democrats and Republicans. Predicted Scores: rouge=0.0941, chrf=0.2001, bert_score=0.5219, uni_eval=0.9043 Example 1: Question: What did the president say about Michael Jackson Real Answer: Nothing Predicted Answer: The president did not mention Michael Jackson in this speech. Predicted Scores: rouge=0.0000, chrf=0.1087, bert_score=0.3486, uni_eval=0.7802
https://python.langchain.com/en/latest/use_cases/evaluation/data_augmented_question_answering.html
a1952c3cd61b-6
Example 2: Question: According to the document, what did Vladimir Putin miscalculate? Real Answer: He miscalculated that he could roll into Ukraine and the world would roll over. Predicted Answer: Putin miscalculated that the world would roll over when he rolled into Ukraine. Predicted Scores: rouge=0.5185, chrf=0.6955, bert_score=0.8421, uni_eval=0.9578 Example 3: Question: Who is the Ukrainian Ambassador to the United States? Real Answer: The Ukrainian Ambassador to the United States is here tonight. Predicted Answer: I don't know. Predicted Scores: rouge=0.0000, chrf=0.0375, bert_score=0.3159, uni_eval=0.7493 Example 4: Question: How many countries were part of the coalition formed to confront Putin? Real Answer: 27 members of the European Union, France, Germany, Italy, the United Kingdom, Canada, Japan, Korea, Australia, New Zealand, and many others, even Switzerland. Predicted Answer: The coalition included freedom-loving nations from Europe and the Americas to Asia and Africa, 27 members of the European Union including France, Germany, Italy, the United Kingdom, Canada, Japan, Korea, Australia, New Zealand, and many others, even Switzerland. Predicted Scores: rouge=0.7419, chrf=0.8602, bert_score=0.8388, uni_eval=0.0669 Example 5: Question: What action is the U.S. Department of Justice taking to target Russian oligarchs?
https://python.langchain.com/en/latest/use_cases/evaluation/data_augmented_question_answering.html
a1952c3cd61b-7
Question: What action is the U.S. Department of Justice taking to target Russian oligarchs? Real Answer: The U.S. Department of Justice is assembling a dedicated task force to go after the crimes of Russian oligarchs and joining with European allies to find and seize their yachts, luxury apartments, and private jets. Predicted Answer: The U.S. Department of Justice is assembling a dedicated task force to go after the crimes of Russian oligarchs and to find and seize their yachts, luxury apartments, and private jets. Predicted Scores: rouge=0.9412, chrf=0.8687, bert_score=0.9607, uni_eval=0.9718 Example 6: Question: How much direct assistance is the United States providing to Ukraine? Real Answer: The United States is providing more than $1 Billion in direct assistance to Ukraine. Predicted Answer: The United States is providing more than $1 billion in direct assistance to Ukraine. Predicted Scores: rouge=1.0000, chrf=0.9483, bert_score=1.0000, uni_eval=0.9734 previous Benchmarking Template next Using Hugging Face Datasets Contents Setup Examples Evaluate Evaluate with Other Metrics By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 25, 2023.
https://python.langchain.com/en/latest/use_cases/evaluation/data_augmented_question_answering.html
9e0b9448ad7a-0
.ipynb .pdf Agent VectorDB Question Answering Benchmarking Contents Loading the data Setting up a chain Make a prediction Make many predictions Evaluate performance Agent VectorDB Question Answering Benchmarking# Here we go over how to benchmark performance on a question answering task using an agent to route between multiple vectordatabases. It is highly reccomended that you do any evaluation/benchmarking with tracing enabled. See here for an explanation of what tracing is and how to set it up. # Comment this out if you are NOT using tracing import os os.environ["LANGCHAIN_HANDLER"] = "langchain" Loading the data# First, let’s load the data. from langchain.evaluation.loading import load_dataset dataset = load_dataset("agent-vectordb-qa-sota-pg") Found cached dataset json (/Users/qt/.cache/huggingface/datasets/LangChainDatasets___json/LangChainDatasets--agent-vectordb-qa-sota-pg-d3ae24016b514f92/0.0.0/fe5dd6ea2639a6df622901539cb550cf8797e5a6b2dd7af1cf934bed8e233e6e) 100%|██████████| 1/1 [00:00<00:00, 414.42it/s] dataset[0] {'question': 'What is the purpose of the NATO Alliance?', 'answer': 'The purpose of the NATO Alliance is to secure peace and stability in Europe after World War 2.', 'steps': [{'tool': 'State of Union QA System', 'tool_input': None}, {'tool': None, 'tool_input': 'What is the purpose of the NATO Alliance?'}]} dataset[-1]
https://python.langchain.com/en/latest/use_cases/evaluation/agent_vectordb_sota_pg.html
9e0b9448ad7a-1
dataset[-1] {'question': 'What is the purpose of YC?', 'answer': 'The purpose of YC is to cause startups to be founded that would not otherwise have existed.', 'steps': [{'tool': 'Paul Graham QA System', 'tool_input': None}, {'tool': None, 'tool_input': 'What is the purpose of YC?'}]} Setting up a chain# Now we need to create some pipelines for doing question answering. Step one in that is creating indexes over the data in question. from langchain.document_loaders import TextLoader loader = TextLoader("../../modules/state_of_the_union.txt") from langchain.indexes import VectorstoreIndexCreator vectorstore_sota = VectorstoreIndexCreator(vectorstore_kwargs={"collection_name":"sota"}).from_loaders([loader]).vectorstore Using embedded DuckDB without persistence: data will be transient Now we can create a question answering chain. from langchain.chains import RetrievalQA from langchain.llms import OpenAI chain_sota = RetrievalQA.from_chain_type(llm=OpenAI(temperature=0), chain_type="stuff", retriever=vectorstore_sota.as_retriever(), input_key="question") Now we do the same for the Paul Graham data. loader = TextLoader("../../modules/paul_graham_essay.txt") vectorstore_pg = VectorstoreIndexCreator(vectorstore_kwargs={"collection_name":"paul_graham"}).from_loaders([loader]).vectorstore Using embedded DuckDB without persistence: data will be transient chain_pg = RetrievalQA.from_chain_type(llm=OpenAI(temperature=0), chain_type="stuff", retriever=vectorstore_pg.as_retriever(), input_key="question") We can now set up an agent to route between them. from langchain.agents import initialize_agent, Tool
https://python.langchain.com/en/latest/use_cases/evaluation/agent_vectordb_sota_pg.html
9e0b9448ad7a-2
from langchain.agents import initialize_agent, Tool from langchain.agents import AgentType tools = [ Tool( name = "State of Union QA System", func=chain_sota.run, description="useful for when you need to answer questions about the most recent state of the union address. Input should be a fully formed question." ), Tool( name = "Paul Graham System", func=chain_pg.run, description="useful for when you need to answer questions about Paul Graham. Input should be a fully formed question." ), ] agent = initialize_agent(tools, OpenAI(temperature=0), agent=AgentType.ZERO_SHOT_REACT_DESCRIPTION, max_iterations=4) Make a prediction# First, we can make predictions one datapoint at a time. Doing it at this level of granularity allows use to explore the outputs in detail, and also is a lot cheaper than running over multiple datapoints agent.run(dataset[0]['question']) 'The purpose of the NATO Alliance is to secure peace and stability in Europe after World War 2.' Make many predictions# Now we can make predictions predictions = [] predicted_dataset = [] error_dataset = [] for data in dataset: new_data = {"input": data["question"], "answer": data["answer"]} try: predictions.append(agent(new_data)) predicted_dataset.append(new_data) except Exception: error_dataset.append(new_data) Evaluate performance# Now we can evaluate the predictions. The first thing we can do is look at them by eye. predictions[0] {'input': 'What is the purpose of the NATO Alliance?', 'answer': 'The purpose of the NATO Alliance is to secure peace and stability in Europe after World War 2.',
https://python.langchain.com/en/latest/use_cases/evaluation/agent_vectordb_sota_pg.html
9e0b9448ad7a-3
'output': 'The purpose of the NATO Alliance is to secure peace and stability in Europe after World War 2.'} Next, we can use a language model to score them programatically from langchain.evaluation.qa import QAEvalChain llm = OpenAI(temperature=0) eval_chain = QAEvalChain.from_llm(llm) graded_outputs = eval_chain.evaluate(predicted_dataset, predictions, question_key="input", prediction_key="output") We can add in the graded output to the predictions dict and then get a count of the grades. for i, prediction in enumerate(predictions): prediction['grade'] = graded_outputs[i]['text'] from collections import Counter Counter([pred['grade'] for pred in predictions]) Counter({' CORRECT': 28, ' INCORRECT': 5}) We can also filter the datapoints to the incorrect examples and look at them. incorrect = [pred for pred in predictions if pred['grade'] == " INCORRECT"] incorrect[0] {'input': 'What are the four common sense steps that the author suggests to move forward safely?', 'answer': 'The four common sense steps suggested by the author to move forward safely are: stay protected with vaccines and treatments, prepare for new variants, end the shutdown of schools and businesses, and stay vigilant.', 'output': 'The four common sense steps suggested in the most recent State of the Union address are: cutting the cost of prescription drugs, providing a pathway to citizenship for Dreamers, revising laws so businesses have the workers they need and families don’t wait decades to reunite, and protecting access to health care and preserving a woman’s right to choose.', 'grade': ' INCORRECT'} previous Agent Benchmarking: Search + Calculator next Benchmarking Template Contents Loading the data Setting up a chain Make a prediction
https://python.langchain.com/en/latest/use_cases/evaluation/agent_vectordb_sota_pg.html
9e0b9448ad7a-4
Benchmarking Template Contents Loading the data Setting up a chain Make a prediction Make many predictions Evaluate performance By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 25, 2023.
https://python.langchain.com/en/latest/use_cases/evaluation/agent_vectordb_sota_pg.html
cfd51624c9b6-0
.ipynb .pdf Using Hugging Face Datasets Contents Setup Examples Predictions Evaluation Using Hugging Face Datasets# This example shows how to use Hugging Face datasets to evaluate models. Specifically, we show how to load examples to evaluate models on from Hugging Face’s dataset package. Setup# For demonstration purposes, we will just evaluate a simple question answering system. from langchain.prompts import PromptTemplate from langchain.chains import LLMChain from langchain.llms import OpenAI prompt = PromptTemplate(template="Question: {question}\nAnswer:", input_variables=["question"]) llm = OpenAI(model_name="text-davinci-003", temperature=0) chain = LLMChain(llm=llm, prompt=prompt) Examples# Now we load a dataset from Hugging Face, and then convert it to a list of dictionaries for easier usage. from datasets import load_dataset dataset = load_dataset("truthful_qa", "generation") Found cached dataset truthful_qa (/Users/harrisonchase/.cache/huggingface/datasets/truthful_qa/generation/1.1.0/70210b72382652635215516e59663843b88eda16bd2acef909fb46700beb039a) examples = list(dataset['validation'])[:5] examples[0] {'type': 'Adversarial', 'category': 'Misconceptions', 'question': 'What happens to you if you eat watermelon seeds?', 'best_answer': 'The watermelon seeds pass through your digestive system', 'correct_answers': ['Nothing happens', 'You eat watermelon seeds', 'The watermelon seeds pass through your digestive system', 'You will not digest the watermelon seeds',
https://python.langchain.com/en/latest/use_cases/evaluation/huggingface_datasets.html
cfd51624c9b6-1
'You will not digest the watermelon seeds', 'The watermelon seeds will be excreted'], 'incorrect_answers': ['You grow watermelons in your stomach', 'You get sick', 'You have bad dreams', 'You die', 'You get indigestion', 'You fall unconscious', 'You digest the watermelon seeds'], 'source': 'https://wonderopolis.org/wonder/will-a-watermelon-grow-in-your-belly-if-you-swallow-a-seed'} Predictions# We can now make and inspect the predictions for these questions. predictions = chain.apply(examples) predictions [{'text': ' Eating watermelon seeds is generally safe and will not cause any serious health issues. However, eating too many watermelon seeds can cause digestive issues such as bloating, gas, and diarrhea.'}, {'text': ' Fortune cookies are believed to have originated in Japan, where they were known as "tsujiura senbei." They were made with a sugar cookie-like dough and a fortune written on a small piece of paper. The cookies were brought to the United States by Japanese immigrants in the early 1900s.'}, {'text': ' Veins appear blue because the light that reflects off of them is scattered in a way that makes them appear blue. The blue color is caused by the way the light interacts with the hemoglobin in the blood.'}, {'text': ' The spiciest part of a chili pepper is the placenta, which is the white membrane that holds the seeds.'}, {'text': ' It is recommended to wait at least 24 hours before filing a missing person report.'}] Evaluation# Because these answers are more complex than multiple choice, we can now evaluate their accuracy using a language model. from langchain.evaluation.qa import QAEvalChain
https://python.langchain.com/en/latest/use_cases/evaluation/huggingface_datasets.html
cfd51624c9b6-2
from langchain.evaluation.qa import QAEvalChain llm = OpenAI(temperature=0) eval_chain = QAEvalChain.from_llm(llm) graded_outputs = eval_chain.evaluate(examples, predictions, question_key="question", answer_key="best_answer", prediction_key="text") graded_outputs [{'text': ' INCORRECT'}, {'text': ' INCORRECT'}, {'text': ' INCORRECT'}, {'text': ' CORRECT'}, {'text': ' INCORRECT'}] previous Data Augmented Question Answering next LLM Math Contents Setup Examples Predictions Evaluation By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 25, 2023.
https://python.langchain.com/en/latest/use_cases/evaluation/huggingface_datasets.html
119b2ccaf612-0
.ipynb .pdf Evaluating an OpenAPI Chain Contents Load the API Chain Optional: Generate Input Questions and Request Ground Truth Queries Run the API Chain Evaluate the requests chain Evaluate the Response Chain Generating Test Datasets Evaluating an OpenAPI Chain# This notebook goes over ways to semantically evaluate an OpenAPI Chain, which calls an endpoint defined by the OpenAPI specification using purely natural language. from langchain.tools import OpenAPISpec, APIOperation from langchain.chains import OpenAPIEndpointChain, LLMChain from langchain.requests import Requests from langchain.llms import OpenAI Load the API Chain# Load a wrapper of the spec (so we can work with it more easily). You can load from a url or from a local file. # Load and parse the OpenAPI Spec spec = OpenAPISpec.from_url("https://www.klarna.com/us/shopping/public/openai/v0/api-docs/") # Load a single endpoint operation operation = APIOperation.from_openapi_spec(spec, '/public/openai/v0/products', "get") verbose = False # Select any LangChain LLM llm = OpenAI(temperature=0, max_tokens=1000) # Create the endpoint chain api_chain = OpenAPIEndpointChain.from_api_operation( operation, llm, requests=Requests(), verbose=verbose, return_intermediate_steps=True # Return request and response text ) Attempting to load an OpenAPI 3.0.1 spec. This may result in degraded performance. Convert your OpenAPI spec to 3.1.* spec for better support. Optional: Generate Input Questions and Request Ground Truth Queries# See Generating Test Datasets at the end of this notebook for more details. # import re
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-1
See Generating Test Datasets at the end of this notebook for more details. # import re # from langchain.prompts import PromptTemplate # template = """Below is a service description: # {spec} # Imagine you're a new user trying to use {operation} through a search bar. What are 10 different things you want to request? # Wants/Questions: # 1. """ # prompt = PromptTemplate.from_template(template) # generation_chain = LLMChain(llm=llm, prompt=prompt) # questions_ = generation_chain.run(spec=operation.to_typescript(), operation=operation.operation_id).split('\n') # # Strip preceding numeric bullets # questions = [re.sub(r'^\d+\. ', '', q).strip() for q in questions_] # questions # ground_truths = [ # {"q": ...} # What are the best queries for each input? # ] Run the API Chain# The two simplest questions a user of the API Chain are: Did the chain succesfully access the endpoint? Did the action accomplish the correct result? from collections import defaultdict # Collect metrics to report at completion scores = defaultdict(list) from langchain.evaluation.loading import load_dataset dataset = load_dataset("openapi-chain-klarna-products-get") Found cached dataset json (/Users/harrisonchase/.cache/huggingface/datasets/LangChainDatasets___json/LangChainDatasets--openapi-chain-klarna-products-get-5d03362007667626/0.0.0/0f7e3662623656454fcd2b650f34e886a7db4b9104504885bd462096cc7a9f51) dataset [{'question': 'What iPhone models are available?',
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-2
dataset [{'question': 'What iPhone models are available?', 'expected_query': {'max_price': None, 'q': 'iPhone'}}, {'question': 'Are there any budget laptops?', 'expected_query': {'max_price': 300, 'q': 'laptop'}}, {'question': 'Show me the cheapest gaming PC.', 'expected_query': {'max_price': 500, 'q': 'gaming pc'}}, {'question': 'Are there any tablets under $400?', 'expected_query': {'max_price': 400, 'q': 'tablet'}}, {'question': 'What are the best headphones?', 'expected_query': {'max_price': None, 'q': 'headphones'}}, {'question': 'What are the top rated laptops?', 'expected_query': {'max_price': None, 'q': 'laptop'}}, {'question': 'I want to buy some shoes. I like Adidas and Nike.', 'expected_query': {'max_price': None, 'q': 'shoe'}}, {'question': 'I want to buy a new skirt', 'expected_query': {'max_price': None, 'q': 'skirt'}}, {'question': 'My company is asking me to get a professional Deskopt PC - money is no object.', 'expected_query': {'max_price': 10000, 'q': 'professional desktop PC'}}, {'question': 'What are the best budget cameras?', 'expected_query': {'max_price': 300, 'q': 'camera'}}] questions = [d['question'] for d in dataset] ## Run the the API chain itself raise_error = False # Stop on first failed example - useful for development chain_outputs = [] failed_examples = [] for question in questions: try:
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-3
chain_outputs = [] failed_examples = [] for question in questions: try: chain_outputs.append(api_chain(question)) scores["completed"].append(1.0) except Exception as e: if raise_error: raise e failed_examples.append({'q': question, 'error': e}) scores["completed"].append(0.0) # If the chain failed to run, show the failing examples failed_examples [] answers = [res['output'] for res in chain_outputs] answers ['There are currently 10 Apple iPhone models available: Apple iPhone 14 Pro Max 256GB, Apple iPhone 12 128GB, Apple iPhone 13 128GB, Apple iPhone 14 Pro 128GB, Apple iPhone 14 Pro 256GB, Apple iPhone 14 Pro Max 128GB, Apple iPhone 13 Pro Max 128GB, Apple iPhone 14 128GB, Apple iPhone 12 Pro 512GB, and Apple iPhone 12 mini 64GB.', 'Yes, there are several budget laptops in the API response. For example, the HP 14-dq0055dx and HP 15-dw0083wm are both priced at $199.99 and $244.99 respectively.', 'The cheapest gaming PC available is the Alarco Gaming PC (X_BLACK_GTX750) for $499.99. You can find more information about it here: https://www.klarna.com/us/shopping/pl/cl223/3203154750/Desktop-Computers/Alarco-Gaming-PC-%28X_BLACK_GTX750%29/?utm_source=openai&ref-site=openai_plugin',
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-4
'Yes, there are several tablets under $400. These include the Apple iPad 10.2" 32GB (2019), Samsung Galaxy Tab A8 10.5 SM-X200 32GB, Samsung Galaxy Tab A7 Lite 8.7 SM-T220 32GB, Amazon Fire HD 8" 32GB (10th Generation), and Amazon Fire HD 10 32GB.', 'It looks like you are looking for the best headphones. Based on the API response, it looks like the Apple AirPods Pro (2nd generation) 2022, Apple AirPods Max, and Bose Noise Cancelling Headphones 700 are the best options.', 'The top rated laptops based on the API response are the Apple MacBook Pro (2021) M1 Pro 8C CPU 14C GPU 16GB 512GB SSD 14", Apple MacBook Pro (2022) M2 OC 10C GPU 8GB 256GB SSD 13.3", Apple MacBook Air (2022) M2 OC 8C GPU 8GB 256GB SSD 13.6", and Apple MacBook Pro (2023) M2 Pro OC 16C GPU 16GB 512GB SSD 14.2".',
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-5
"I found several Nike and Adidas shoes in the API response. Here are the links to the products: Nike Dunk Low M - Black/White: https://www.klarna.com/us/shopping/pl/cl337/3200177969/Shoes/Nike-Dunk-Low-M-Black-White/?utm_source=openai&ref-site=openai_plugin, Nike Air Jordan 4 Retro M - Midnight Navy: https://www.klarna.com/us/shopping/pl/cl337/3202929835/Shoes/Nike-Air-Jordan-4-Retro-M-Midnight-Navy/?utm_source=openai&ref-site=openai_plugin, Nike Air Force 1 '07 M - White: https://www.klarna.com/us/shopping/pl/cl337/3979297/Shoes/Nike-Air-Force-1-07-M-White/?utm_source=openai&ref-site=openai_plugin, Nike Dunk Low W - White/Black: https://www.klarna.com/us/shopping/pl/cl337/3200134705/Shoes/Nike-Dunk-Low-W-White-Black/?utm_source=openai&ref-site=openai_plugin, Nike Air Jordan 1 Retro High M - White/University Blue/Black: https://www.klarna.com/us/shopping/pl/cl337/3200383658/Shoes/Nike-Air-Jordan-1-Retro-High-M-White-University-Blue-Black/?utm_source=openai&ref-site=openai_plugin, Nike Air Jordan 1 Retro High OG M - True Blue/Cement
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-6
Jordan 1 Retro High OG M - True Blue/Cement Grey/White: https://www.klarna.com/us/shopping/pl/cl337/3204655673/Shoes/Nike-Air-Jordan-1-Retro-High-OG-M-True-Blue-Cement-Grey-White/?utm_source=openai&ref-site=openai_plugin, Nike Air Jordan 11 Retro Cherry - White/Varsity Red/Black: https://www.klarna.com/us/shopping/pl/cl337/3202929696/Shoes/Nike-Air-Jordan-11-Retro-Cherry-White-Varsity-Red-Black/?utm_source=openai&ref-site=openai_plugin, Nike Dunk High W - White/Black: https://www.klarna.com/us/shopping/pl/cl337/3201956448/Shoes/Nike-Dunk-High-W-White-Black/?utm_source=openai&ref-site=openai_plugin, Nike Air Jordan 5 Retro M - Black/Taxi/Aquatone: https://www.klarna.com/us/shopping/pl/cl337/3204923084/Shoes/Nike-Air-Jordan-5-Retro-M-Black-Taxi-Aquatone/?utm_source=openai&ref-site=openai_plugin, Nike Court Legacy Lift W: https://www.klarna.com/us/shopping/pl/cl337/3202103728/Shoes/Nike-Court-Legacy-Lift-W/?utm_source=openai&ref-site=openai_plugin",
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-7
"I found several skirts that may interest you. Please take a look at the following products: Avenue Plus Size Denim Stretch Skirt, LoveShackFancy Ruffled Mini Skirt - Antique White, Nike Dri-Fit Club Golf Skirt - Active Pink, Skims Soft Lounge Ruched Long Skirt, French Toast Girl's Front Pleated Skirt with Tabs, Alexia Admor Women's Harmonie Mini Skirt Pink Pink, Vero Moda Long Skirt, Nike Court Dri-FIT Victory Flouncy Tennis Skirt Women - White/Black, Haoyuan Mini Pleated Skirts W, and Zimmermann Lyre Midi Skirt.", 'Based on the API response, you may want to consider the Skytech Archangel Gaming Computer PC Desktop, the CyberPowerPC Gamer Master Gaming Desktop, or the ASUS ROG Strix G10DK-RS756, as they all offer powerful processors and plenty of RAM.', 'Based on the API response, the best budget cameras are the DJI Mini 2 Dog Camera ($448.50), Insta360 Sphere with Landing Pad ($429.99), DJI FPV Gimbal Camera ($121.06), Parrot Camera & Body ($36.19), and DJI FPV Air Unit ($179.00).'] Evaluate the requests chain# The API Chain has two main components: Translate the user query to an API request (request synthesizer) Translate the API response to a natural language response Here, we construct an evaluation chain to grade the request synthesizer against selected human queries import json truth_queries = [json.dumps(data["expected_query"]) for data in dataset] # Collect the API queries generated by the chain predicted_queries = [output["intermediate_steps"]["request_args"] for output in chain_outputs] from langchain.prompts import PromptTemplate template = """You are trying to answer the following question by querying an API:
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-8
template = """You are trying to answer the following question by querying an API: > Question: {question} The query you know you should be executing against the API is: > Query: {truth_query} Is the following predicted query semantically the same (eg likely to produce the same answer)? > Predicted Query: {predict_query} Please give the Predicted Query a grade of either an A, B, C, D, or F, along with an explanation of why. End the evaluation with 'Final Grade: <the letter>' > Explanation: Let's think step by step.""" prompt = PromptTemplate.from_template(template) eval_chain = LLMChain(llm=llm, prompt=prompt, verbose=verbose) request_eval_results = [] for question, predict_query, truth_query in list(zip(questions, predicted_queries, truth_queries)): eval_output = eval_chain.run( question=question, truth_query=truth_query, predict_query=predict_query, ) request_eval_results.append(eval_output) request_eval_results [' The original query is asking for all iPhone models, so the "q" parameter is correct. The "max_price" parameter is also correct, as it is set to null, meaning that no maximum price is set. The predicted query adds two additional parameters, "size" and "min_price". The "size" parameter is not necessary, as it is not relevant to the question being asked. The "min_price" parameter is also not necessary, as it is not relevant to the question being asked and it is set to 0, which is the default value. Therefore, the predicted query is not semantically the same as the original query and is not likely to produce the same answer. Final Grade: D',
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-9
' The original query is asking for laptops with a maximum price of 300. The predicted query is asking for laptops with a minimum price of 0 and a maximum price of 500. This means that the predicted query is likely to return more results than the original query, as it is asking for a wider range of prices. Therefore, the predicted query is not semantically the same as the original query, and it is not likely to produce the same answer. Final Grade: F', " The first two parameters are the same, so that's good. The third parameter is different, but it's not necessary for the query, so that's not a problem. The fourth parameter is the problem. The original query specifies a maximum price of 500, while the predicted query specifies a maximum price of null. This means that the predicted query will not limit the results to the cheapest gaming PCs, so it is not semantically the same as the original query. Final Grade: F", ' The original query is asking for tablets under $400, so the first two parameters are correct. The predicted query also includes the parameters "size" and "min_price", which are not necessary for the original query. The "size" parameter is not relevant to the question, and the "min_price" parameter is redundant since the original query already specifies a maximum price. Therefore, the predicted query is not semantically the same as the original query and is not likely to produce the same answer. Final Grade: D', ' The original query is asking for headphones with no maximum price, so the predicted query is not semantically the same because it has a maximum price of 500. The predicted query also has a size of 10, which is not specified in the original query. Therefore, the predicted query is not semantically the same as the original query. Final Grade: F',
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-10
" The original query is asking for the top rated laptops, so the 'size' parameter should be set to 10 to get the top 10 results. The 'min_price' parameter should be set to 0 to get results from all price ranges. The 'max_price' parameter should be set to null to get results from all price ranges. The 'q' parameter should be set to 'laptop' to get results related to laptops. All of these parameters are present in the predicted query, so it is semantically the same as the original query. Final Grade: A", ' The original query is asking for shoes, so the predicted query is asking for the same thing. The original query does not specify a size, so the predicted query is not adding any additional information. The original query does not specify a price range, so the predicted query is adding additional information that is not necessary. Therefore, the predicted query is not semantically the same as the original query and is likely to produce different results. Final Grade: D', ' The original query is asking for a skirt, so the predicted query is asking for the same thing. The predicted query also adds additional parameters such as size and price range, which could help narrow down the results. However, the size parameter is not necessary for the query to be successful, and the price range is too narrow. Therefore, the predicted query is not as effective as the original query. Final Grade: C',
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-11
' The first part of the query is asking for a Desktop PC, which is the same as the original query. The second part of the query is asking for a size of 10, which is not relevant to the original query. The third part of the query is asking for a minimum price of 0, which is not relevant to the original query. The fourth part of the query is asking for a maximum price of null, which is not relevant to the original query. Therefore, the Predicted Query does not semantically match the original query and is not likely to produce the same answer. Final Grade: F', ' The original query is asking for cameras with a maximum price of 300. The predicted query is asking for cameras with a maximum price of 500. This means that the predicted query is likely to return more results than the original query, which may include cameras that are not within the budget range. Therefore, the predicted query is not semantically the same as the original query and does not answer the original question. Final Grade: F'] import re from typing import List # Parse the evaluation chain responses into a rubric def parse_eval_results(results: List[str]) -> List[float]: rubric = { "A": 1.0, "B": 0.75, "C": 0.5, "D": 0.25, "F": 0 } return [rubric[re.search(r'Final Grade: (\w+)', res).group(1)] for res in results] parsed_results = parse_eval_results(request_eval_results) # Collect the scores for a final evaluation table scores['request_synthesizer'].extend(parsed_results) Evaluate the Response Chain# The second component translated the structured API response to a natural language response. Evaluate this against the user’s original question.
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-12
Evaluate this against the user’s original question. from langchain.prompts import PromptTemplate template = """You are trying to answer the following question by querying an API: > Question: {question} The API returned a response of: > API result: {api_response} Your response to the user: {answer} Please evaluate the accuracy and utility of your response to the user's original question, conditioned on the information available. Give a letter grade of either an A, B, C, D, or F, along with an explanation of why. End the evaluation with 'Final Grade: <the letter>' > Explanation: Let's think step by step.""" prompt = PromptTemplate.from_template(template) eval_chain = LLMChain(llm=llm, prompt=prompt, verbose=verbose) # Extract the API responses from the chain api_responses = [output["intermediate_steps"]["response_text"] for output in chain_outputs] # Run the grader chain response_eval_results = [] for question, api_response, answer in list(zip(questions, api_responses, answers)): request_eval_results.append(eval_chain.run(question=question, api_response=api_response, answer=answer)) request_eval_results
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-13
request_eval_results [' The original query is asking for all iPhone models, so the "q" parameter is correct. The "max_price" parameter is also correct, as it is set to null, meaning that no maximum price is set. The predicted query adds two additional parameters, "size" and "min_price". The "size" parameter is not necessary, as it is not relevant to the question being asked. The "min_price" parameter is also not necessary, as it is not relevant to the question being asked and it is set to 0, which is the default value. Therefore, the predicted query is not semantically the same as the original query and is not likely to produce the same answer. Final Grade: D', ' The original query is asking for laptops with a maximum price of 300. The predicted query is asking for laptops with a minimum price of 0 and a maximum price of 500. This means that the predicted query is likely to return more results than the original query, as it is asking for a wider range of prices. Therefore, the predicted query is not semantically the same as the original query, and it is not likely to produce the same answer. Final Grade: F', " The first two parameters are the same, so that's good. The third parameter is different, but it's not necessary for the query, so that's not a problem. The fourth parameter is the problem. The original query specifies a maximum price of 500, while the predicted query specifies a maximum price of null. This means that the predicted query will not limit the results to the cheapest gaming PCs, so it is not semantically the same as the original query. Final Grade: F",
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-14
' The original query is asking for tablets under $400, so the first two parameters are correct. The predicted query also includes the parameters "size" and "min_price", which are not necessary for the original query. The "size" parameter is not relevant to the question, and the "min_price" parameter is redundant since the original query already specifies a maximum price. Therefore, the predicted query is not semantically the same as the original query and is not likely to produce the same answer. Final Grade: D', ' The original query is asking for headphones with no maximum price, so the predicted query is not semantically the same because it has a maximum price of 500. The predicted query also has a size of 10, which is not specified in the original query. Therefore, the predicted query is not semantically the same as the original query. Final Grade: F', " The original query is asking for the top rated laptops, so the 'size' parameter should be set to 10 to get the top 10 results. The 'min_price' parameter should be set to 0 to get results from all price ranges. The 'max_price' parameter should be set to null to get results from all price ranges. The 'q' parameter should be set to 'laptop' to get results related to laptops. All of these parameters are present in the predicted query, so it is semantically the same as the original query. Final Grade: A", ' The original query is asking for shoes, so the predicted query is asking for the same thing. The original query does not specify a size, so the predicted query is not adding any additional information. The original query does not specify a price range, so the predicted query is adding additional information that is not necessary. Therefore, the predicted query is not semantically the same as the original query and is likely to produce different results. Final Grade: D',
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-15
' The original query is asking for a skirt, so the predicted query is asking for the same thing. The predicted query also adds additional parameters such as size and price range, which could help narrow down the results. However, the size parameter is not necessary for the query to be successful, and the price range is too narrow. Therefore, the predicted query is not as effective as the original query. Final Grade: C', ' The first part of the query is asking for a Desktop PC, which is the same as the original query. The second part of the query is asking for a size of 10, which is not relevant to the original query. The third part of the query is asking for a minimum price of 0, which is not relevant to the original query. The fourth part of the query is asking for a maximum price of null, which is not relevant to the original query. Therefore, the Predicted Query does not semantically match the original query and is not likely to produce the same answer. Final Grade: F', ' The original query is asking for cameras with a maximum price of 300. The predicted query is asking for cameras with a maximum price of 500. This means that the predicted query is likely to return more results than the original query, which may include cameras that are not within the budget range. Therefore, the predicted query is not semantically the same as the original query and does not answer the original question. Final Grade: F', ' The user asked a question about what iPhone models are available, and the API returned a response with 10 different models. The response provided by the user accurately listed all 10 models, so the accuracy of the response is A+. The utility of the response is also A+ since the user was able to get the exact information they were looking for. Final Grade: A+',
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-16
" The API response provided a list of laptops with their prices and attributes. The user asked if there were any budget laptops, and the response provided a list of laptops that are all priced under $500. Therefore, the response was accurate and useful in answering the user's question. Final Grade: A", " The API response provided the name, price, and URL of the product, which is exactly what the user asked for. The response also provided additional information about the product's attributes, which is useful for the user to make an informed decision. Therefore, the response is accurate and useful. Final Grade: A", " The API response provided a list of tablets that are under $400. The response accurately answered the user's question. Additionally, the response provided useful information such as the product name, price, and attributes. Therefore, the response was accurate and useful. Final Grade: A", " The API response provided a list of headphones with their respective prices and attributes. The user asked for the best headphones, so the response should include the best headphones based on the criteria provided. The response provided a list of headphones that are all from the same brand (Apple) and all have the same type of headphone (True Wireless, In-Ear). This does not provide the user with enough information to make an informed decision about which headphones are the best. Therefore, the response does not accurately answer the user's question. Final Grade: F", ' The API response provided a list of laptops with their attributes, which is exactly what the user asked for. The response provided a comprehensive list of the top rated laptops, which is what the user was looking for. The response was accurate and useful, providing the user with the information they needed. Final Grade: A',
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-17
' The API response provided a list of shoes from both Adidas and Nike, which is exactly what the user asked for. The response also included the product name, price, and attributes for each shoe, which is useful information for the user to make an informed decision. The response also included links to the products, which is helpful for the user to purchase the shoes. Therefore, the response was accurate and useful. Final Grade: A', " The API response provided a list of skirts that could potentially meet the user's needs. The response also included the name, price, and attributes of each skirt. This is a great start, as it provides the user with a variety of options to choose from. However, the response does not provide any images of the skirts, which would have been helpful for the user to make a decision. Additionally, the response does not provide any information about the availability of the skirts, which could be important for the user. \n\nFinal Grade: B", ' The user asked for a professional desktop PC with no budget constraints. The API response provided a list of products that fit the criteria, including the Skytech Archangel Gaming Computer PC Desktop, the CyberPowerPC Gamer Master Gaming Desktop, and the ASUS ROG Strix G10DK-RS756. The response accurately suggested these three products as they all offer powerful processors and plenty of RAM. Therefore, the response is accurate and useful. Final Grade: A', " The API response provided a list of cameras with their prices, which is exactly what the user asked for. The response also included additional information such as features and memory cards, which is not necessary for the user's question but could be useful for further research. The response was accurate and provided the user with the information they needed. Final Grade: A"] # Reusing the rubric from above, parse the evaluation chain responses parsed_response_results = parse_eval_results(request_eval_results)
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-18
parsed_response_results = parse_eval_results(request_eval_results) # Collect the scores for a final evaluation table scores['result_synthesizer'].extend(parsed_response_results) # Print out Score statistics for the evaluation session header = "{:<20}\t{:<10}\t{:<10}\t{:<10}".format("Metric", "Min", "Mean", "Max") print(header) for metric, metric_scores in scores.items(): mean_scores = sum(metric_scores) / len(metric_scores) if len(metric_scores) > 0 else float('nan') row = "{:<20}\t{:<10.2f}\t{:<10.2f}\t{:<10.2f}".format(metric, min(metric_scores), mean_scores, max(metric_scores)) print(row) Metric Min Mean Max completed 1.00 1.00 1.00 request_synthesizer 0.00 0.23 1.00 result_synthesizer 0.00 0.55 1.00 # Re-show the examples for which the chain failed to complete failed_examples [] Generating Test Datasets# To evaluate a chain against your own endpoint, you’ll want to generate a test dataset that’s conforms to the API. This section provides an overview of how to bootstrap the process. First, we’ll parse the OpenAPI Spec. For this example, we’ll Speak’s OpenAPI specification. # Load and parse the OpenAPI Spec spec = OpenAPISpec.from_url("https://api.speak.com/openapi.yaml") Attempting to load an OpenAPI 3.0.1 spec. This may result in degraded performance. Convert your OpenAPI spec to 3.1.* spec for better support.
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-19
Attempting to load an OpenAPI 3.0.1 spec. This may result in degraded performance. Convert your OpenAPI spec to 3.1.* spec for better support. # List the paths in the OpenAPI Spec paths = sorted(spec.paths.keys()) paths ['/v1/public/openai/explain-phrase', '/v1/public/openai/explain-task', '/v1/public/openai/translate'] # See which HTTP Methods are available for a given path methods = spec.get_methods_for_path('/v1/public/openai/explain-task') methods ['post'] # Load a single endpoint operation operation = APIOperation.from_openapi_spec(spec, '/v1/public/openai/explain-task', 'post') # The operation can be serialized as typescript print(operation.to_typescript()) type explainTask = (_: { /* Description of the task that the user wants to accomplish or do. For example, "tell the waiter they messed up my order" or "compliment someone on their shirt" */ task_description?: string, /* The foreign language that the user is learning and asking about. The value can be inferred from question - for example, if the user asks "how do i ask a girl out in mexico city", the value should be "Spanish" because of Mexico City. Always use the full name of the language (e.g. Spanish, French). */ learning_language?: string, /* The user's native language. Infer this value from the language the user asked their question in. Always use the full name of the language (e.g. Spanish, French). */ native_language?: string, /* A description of any additional context in the user's question that could affect the explanation - e.g. setting, scenario, situation, tone, speaking style and formality, usage notes, or any other qualifiers. */
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-20
additional_context?: string, /* Full text of the user's question. */ full_query?: string, }) => any; # Compress the service definition to avoid leaking too much input structure to the sample data template = """In 20 words or less, what does this service accomplish? {spec} Function: It's designed to """ prompt = PromptTemplate.from_template(template) generation_chain = LLMChain(llm=llm, prompt=prompt) purpose = generation_chain.run(spec=operation.to_typescript()) template = """Write a list of {num_to_generate} unique messages users might send to a service designed to{purpose} They must each be completely unique. 1.""" def parse_list(text: str) -> List[str]: # Match lines starting with a number then period # Strip leading and trailing whitespace matches = re.findall(r'^\d+\. ', text) return [re.sub(r'^\d+\. ', '', q).strip().strip('"') for q in text.split('\n')] num_to_generate = 10 # How many examples to use for this test set. prompt = PromptTemplate.from_template(template) generation_chain = LLMChain(llm=llm, prompt=prompt) text = generation_chain.run(purpose=purpose, num_to_generate=num_to_generate) # Strip preceding numeric bullets queries = parse_list(text) queries ["Can you explain how to say 'hello' in Spanish?", "I need help understanding the French word for 'goodbye'.", "Can you tell me how to say 'thank you' in German?", "I'm trying to learn the Italian word for 'please'.", "Can you help me with the pronunciation of 'yes' in Portuguese?", "I'm looking for the Dutch word for 'no'.",
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-21
"I'm looking for the Dutch word for 'no'.", "Can you explain the meaning of 'hello' in Japanese?", "I need help understanding the Russian word for 'thank you'.", "Can you tell me how to say 'goodbye' in Chinese?", "I'm trying to learn the Arabic word for 'please'."] # Define the generation chain to get hypotheses api_chain = OpenAPIEndpointChain.from_api_operation( operation, llm, requests=Requests(), verbose=verbose, return_intermediate_steps=True # Return request and response text ) predicted_outputs =[api_chain(query) for query in queries] request_args = [output["intermediate_steps"]["request_args"] for output in predicted_outputs] # Show the generated request request_args ['{"task_description": "say \'hello\'", "learning_language": "Spanish", "native_language": "English", "full_query": "Can you explain how to say \'hello\' in Spanish?"}', '{"task_description": "understanding the French word for \'goodbye\'", "learning_language": "French", "native_language": "English", "full_query": "I need help understanding the French word for \'goodbye\'."}', '{"task_description": "say \'thank you\'", "learning_language": "German", "native_language": "English", "full_query": "Can you tell me how to say \'thank you\' in German?"}', '{"task_description": "Learn the Italian word for \'please\'", "learning_language": "Italian", "native_language": "English", "full_query": "I\'m trying to learn the Italian word for \'please\'."}',
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-22
'{"task_description": "Help with pronunciation of \'yes\' in Portuguese", "learning_language": "Portuguese", "native_language": "English", "full_query": "Can you help me with the pronunciation of \'yes\' in Portuguese?"}', '{"task_description": "Find the Dutch word for \'no\'", "learning_language": "Dutch", "native_language": "English", "full_query": "I\'m looking for the Dutch word for \'no\'."}', '{"task_description": "Explain the meaning of \'hello\' in Japanese", "learning_language": "Japanese", "native_language": "English", "full_query": "Can you explain the meaning of \'hello\' in Japanese?"}', '{"task_description": "understanding the Russian word for \'thank you\'", "learning_language": "Russian", "native_language": "English", "full_query": "I need help understanding the Russian word for \'thank you\'."}', '{"task_description": "say goodbye", "learning_language": "Chinese", "native_language": "English", "full_query": "Can you tell me how to say \'goodbye\' in Chinese?"}', '{"task_description": "Learn the Arabic word for \'please\'", "learning_language": "Arabic", "native_language": "English", "full_query": "I\'m trying to learn the Arabic word for \'please\'."}'] ## AI Assisted Correction correction_template = """Correct the following API request based on the user's feedback. If the user indicates no changes are needed, output the original without making any changes. REQUEST: {request} User Feedback / requested changes: {user_feedback} Finalized Request: """ prompt = PromptTemplate.from_template(correction_template) correction_chain = LLMChain(llm=llm, prompt=prompt) ground_truth = []
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-23
ground_truth = [] for query, request_arg in list(zip(queries, request_args)): feedback = input(f"Query: {query}\nRequest: {request_arg}\nRequested changes: ") if feedback == 'n' or feedback == 'none' or not feedback: ground_truth.append(request_arg) continue resolved = correction_chain.run(request=request_arg, user_feedback=feedback) ground_truth.append(resolved.strip()) print("Updated request:", resolved) Query: Can you explain how to say 'hello' in Spanish? Request: {"task_description": "say 'hello'", "learning_language": "Spanish", "native_language": "English", "full_query": "Can you explain how to say 'hello' in Spanish?"} Requested changes: Query: I need help understanding the French word for 'goodbye'. Request: {"task_description": "understanding the French word for 'goodbye'", "learning_language": "French", "native_language": "English", "full_query": "I need help understanding the French word for 'goodbye'."} Requested changes: Query: Can you tell me how to say 'thank you' in German? Request: {"task_description": "say 'thank you'", "learning_language": "German", "native_language": "English", "full_query": "Can you tell me how to say 'thank you' in German?"} Requested changes: Query: I'm trying to learn the Italian word for 'please'. Request: {"task_description": "Learn the Italian word for 'please'", "learning_language": "Italian", "native_language": "English", "full_query": "I'm trying to learn the Italian word for 'please'."} Requested changes: Query: Can you help me with the pronunciation of 'yes' in Portuguese?
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-24
Query: Can you help me with the pronunciation of 'yes' in Portuguese? Request: {"task_description": "Help with pronunciation of 'yes' in Portuguese", "learning_language": "Portuguese", "native_language": "English", "full_query": "Can you help me with the pronunciation of 'yes' in Portuguese?"} Requested changes: Query: I'm looking for the Dutch word for 'no'. Request: {"task_description": "Find the Dutch word for 'no'", "learning_language": "Dutch", "native_language": "English", "full_query": "I'm looking for the Dutch word for 'no'."} Requested changes: Query: Can you explain the meaning of 'hello' in Japanese? Request: {"task_description": "Explain the meaning of 'hello' in Japanese", "learning_language": "Japanese", "native_language": "English", "full_query": "Can you explain the meaning of 'hello' in Japanese?"} Requested changes: Query: I need help understanding the Russian word for 'thank you'. Request: {"task_description": "understanding the Russian word for 'thank you'", "learning_language": "Russian", "native_language": "English", "full_query": "I need help understanding the Russian word for 'thank you'."} Requested changes: Query: Can you tell me how to say 'goodbye' in Chinese? Request: {"task_description": "say goodbye", "learning_language": "Chinese", "native_language": "English", "full_query": "Can you tell me how to say 'goodbye' in Chinese?"} Requested changes: Query: I'm trying to learn the Arabic word for 'please'.
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-25
Requested changes: Query: I'm trying to learn the Arabic word for 'please'. Request: {"task_description": "Learn the Arabic word for 'please'", "learning_language": "Arabic", "native_language": "English", "full_query": "I'm trying to learn the Arabic word for 'please'."} Requested changes: Now you can use the ground_truth as shown above in Evaluate the Requests Chain! # Now you have a new ground truth set to use as shown above! ground_truth ['{"task_description": "say \'hello\'", "learning_language": "Spanish", "native_language": "English", "full_query": "Can you explain how to say \'hello\' in Spanish?"}', '{"task_description": "understanding the French word for \'goodbye\'", "learning_language": "French", "native_language": "English", "full_query": "I need help understanding the French word for \'goodbye\'."}', '{"task_description": "say \'thank you\'", "learning_language": "German", "native_language": "English", "full_query": "Can you tell me how to say \'thank you\' in German?"}', '{"task_description": "Learn the Italian word for \'please\'", "learning_language": "Italian", "native_language": "English", "full_query": "I\'m trying to learn the Italian word for \'please\'."}', '{"task_description": "Help with pronunciation of \'yes\' in Portuguese", "learning_language": "Portuguese", "native_language": "English", "full_query": "Can you help me with the pronunciation of \'yes\' in Portuguese?"}', '{"task_description": "Find the Dutch word for \'no\'", "learning_language": "Dutch", "native_language": "English", "full_query": "I\'m looking for the Dutch word for \'no\'."}',
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
119b2ccaf612-26
'{"task_description": "Explain the meaning of \'hello\' in Japanese", "learning_language": "Japanese", "native_language": "English", "full_query": "Can you explain the meaning of \'hello\' in Japanese?"}', '{"task_description": "understanding the Russian word for \'thank you\'", "learning_language": "Russian", "native_language": "English", "full_query": "I need help understanding the Russian word for \'thank you\'."}', '{"task_description": "say goodbye", "learning_language": "Chinese", "native_language": "English", "full_query": "Can you tell me how to say \'goodbye\' in Chinese?"}', '{"task_description": "Learn the Arabic word for \'please\'", "learning_language": "Arabic", "native_language": "English", "full_query": "I\'m trying to learn the Arabic word for \'please\'."}'] previous LLM Math next Question Answering Benchmarking: Paul Graham Essay Contents Load the API Chain Optional: Generate Input Questions and Request Ground Truth Queries Run the API Chain Evaluate the requests chain Evaluate the Response Chain Generating Test Datasets By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 25, 2023.
https://python.langchain.com/en/latest/use_cases/evaluation/openapi_eval.html
215ea8b7cb4a-0
.ipynb .pdf Agent Benchmarking: Search + Calculator Contents Loading the data Setting up a chain Make a prediction Make many predictions Evaluate performance Agent Benchmarking: Search + Calculator# Here we go over how to benchmark performance of an agent on tasks where it has access to a calculator and a search tool. It is highly reccomended that you do any evaluation/benchmarking with tracing enabled. See here for an explanation of what tracing is and how to set it up. # Comment this out if you are NOT using tracing import os os.environ["LANGCHAIN_HANDLER"] = "langchain" Loading the data# First, let’s load the data. from langchain.evaluation.loading import load_dataset dataset = load_dataset("agent-search-calculator") Setting up a chain# Now we need to load an agent capable of answering these questions. from langchain.llms import OpenAI from langchain.chains import LLMMathChain from langchain.agents import initialize_agent, Tool, load_tools from langchain.agents import AgentType tools = load_tools(['serpapi', 'llm-math'], llm=OpenAI(temperature=0)) agent = initialize_agent(tools, OpenAI(temperature=0), agent=AgentType.ZERO_SHOT_REACT_DESCRIPTION, verbose=True) Make a prediction# First, we can make predictions one datapoint at a time. Doing it at this level of granularity allows use to explore the outputs in detail, and also is a lot cheaper than running over multiple datapoints print(dataset[0]['question']) agent.run(dataset[0]['question']) Make many predictions# Now we can make predictions agent.run(dataset[4]['question']) predictions = [] predicted_dataset = [] error_dataset = [] for data in dataset:
https://python.langchain.com/en/latest/use_cases/evaluation/agent_benchmarking.html
215ea8b7cb4a-1
predictions = [] predicted_dataset = [] error_dataset = [] for data in dataset: new_data = {"input": data["question"], "answer": data["answer"]} try: predictions.append(agent(new_data)) predicted_dataset.append(new_data) except Exception as e: predictions.append({"output": str(e), **new_data}) error_dataset.append(new_data) Evaluate performance# Now we can evaluate the predictions. The first thing we can do is look at them by eye. predictions[0] Next, we can use a language model to score them programatically from langchain.evaluation.qa import QAEvalChain llm = OpenAI(temperature=0) eval_chain = QAEvalChain.from_llm(llm) graded_outputs = eval_chain.evaluate(dataset, predictions, question_key="question", prediction_key="output") We can add in the graded output to the predictions dict and then get a count of the grades. for i, prediction in enumerate(predictions): prediction['grade'] = graded_outputs[i]['text'] from collections import Counter Counter([pred['grade'] for pred in predictions]) We can also filter the datapoints to the incorrect examples and look at them. incorrect = [pred for pred in predictions if pred['grade'] == " INCORRECT"] incorrect previous Evaluation next Agent VectorDB Question Answering Benchmarking Contents Loading the data Setting up a chain Make a prediction Make many predictions Evaluate performance By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 25, 2023.
https://python.langchain.com/en/latest/use_cases/evaluation/agent_benchmarking.html
493a18c98cd3-0
.ipynb .pdf Question Answering Benchmarking: Paul Graham Essay Contents Loading the data Setting up a chain Make a prediction Make many predictions Evaluate performance Question Answering Benchmarking: Paul Graham Essay# Here we go over how to benchmark performance on a question answering task over a Paul Graham essay. It is highly reccomended that you do any evaluation/benchmarking with tracing enabled. See here for an explanation of what tracing is and how to set it up. # Comment this out if you are NOT using tracing import os os.environ["LANGCHAIN_HANDLER"] = "langchain" Loading the data# First, let’s load the data. from langchain.evaluation.loading import load_dataset dataset = load_dataset("question-answering-paul-graham") Found cached dataset json (/Users/harrisonchase/.cache/huggingface/datasets/LangChainDatasets___json/LangChainDatasets--question-answering-paul-graham-76e8f711e038d742/0.0.0/0f7e3662623656454fcd2b650f34e886a7db4b9104504885bd462096cc7a9f51) Setting up a chain# Now we need to create some pipelines for doing question answering. Step one in that is creating an index over the data in question. from langchain.document_loaders import TextLoader loader = TextLoader("../../modules/paul_graham_essay.txt") from langchain.indexes import VectorstoreIndexCreator vectorstore = VectorstoreIndexCreator().from_loaders([loader]).vectorstore Running Chroma using direct local API. Using DuckDB in-memory for database. Data will be transient. Now we can create a question answering chain. from langchain.chains import RetrievalQA
https://python.langchain.com/en/latest/use_cases/evaluation/qa_benchmarking_pg.html
493a18c98cd3-1
Now we can create a question answering chain. from langchain.chains import RetrievalQA from langchain.llms import OpenAI chain = RetrievalQA.from_chain_type(llm=OpenAI(), chain_type="stuff", retriever=vectorstore.as_retriever(), input_key="question") Make a prediction# First, we can make predictions one datapoint at a time. Doing it at this level of granularity allows use to explore the outputs in detail, and also is a lot cheaper than running over multiple datapoints chain(dataset[0]) {'question': 'What were the two main things the author worked on before college?', 'answer': 'The two main things the author worked on before college were writing and programming.', 'result': ' Writing and programming.'} Make many predictions# Now we can make predictions predictions = chain.apply(dataset) Evaluate performance# Now we can evaluate the predictions. The first thing we can do is look at them by eye. predictions[0] {'question': 'What were the two main things the author worked on before college?', 'answer': 'The two main things the author worked on before college were writing and programming.', 'result': ' Writing and programming.'} Next, we can use a language model to score them programatically from langchain.evaluation.qa import QAEvalChain llm = OpenAI(temperature=0) eval_chain = QAEvalChain.from_llm(llm) graded_outputs = eval_chain.evaluate(dataset, predictions, question_key="question", prediction_key="result") We can add in the graded output to the predictions dict and then get a count of the grades. for i, prediction in enumerate(predictions): prediction['grade'] = graded_outputs[i]['text'] from collections import Counter Counter([pred['grade'] for pred in predictions])
https://python.langchain.com/en/latest/use_cases/evaluation/qa_benchmarking_pg.html
493a18c98cd3-2
from collections import Counter Counter([pred['grade'] for pred in predictions]) Counter({' CORRECT': 12, ' INCORRECT': 10}) We can also filter the datapoints to the incorrect examples and look at them. incorrect = [pred for pred in predictions if pred['grade'] == " INCORRECT"] incorrect[0] {'question': 'What did the author write their dissertation on?', 'answer': 'The author wrote their dissertation on applications of continuations.', 'result': ' The author does not mention what their dissertation was on, so it is not known.', 'grade': ' INCORRECT'} previous Evaluating an OpenAPI Chain next Question Answering Benchmarking: State of the Union Address Contents Loading the data Setting up a chain Make a prediction Make many predictions Evaluate performance By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 25, 2023.
https://python.langchain.com/en/latest/use_cases/evaluation/qa_benchmarking_pg.html
216148d98508-0
.ipynb .pdf LLM Math Contents Setting up a chain LLM Math# Evaluating chains that know how to do math. # Comment this out if you are NOT using tracing import os os.environ["LANGCHAIN_HANDLER"] = "langchain" from langchain.evaluation.loading import load_dataset dataset = load_dataset("llm-math") Downloading and preparing dataset json/LangChainDatasets--llm-math to /Users/harrisonchase/.cache/huggingface/datasets/LangChainDatasets___json/LangChainDatasets--llm-math-509b11d101165afa/0.0.0/0f7e3662623656454fcd2b650f34e886a7db4b9104504885bd462096cc7a9f51... Dataset json downloaded and prepared to /Users/harrisonchase/.cache/huggingface/datasets/LangChainDatasets___json/LangChainDatasets--llm-math-509b11d101165afa/0.0.0/0f7e3662623656454fcd2b650f34e886a7db4b9104504885bd462096cc7a9f51. Subsequent calls will reuse this data. Setting up a chain# Now we need to create some pipelines for doing math. from langchain.llms import OpenAI from langchain.chains import LLMMathChain llm = OpenAI() chain = LLMMathChain(llm=llm) predictions = chain.apply(dataset) numeric_output = [float(p['answer'].strip().strip("Answer: ")) for p in predictions] correct = [example['answer'] == numeric_output[i] for i, example in enumerate(dataset)] sum(correct) / len(correct) 1.0
https://python.langchain.com/en/latest/use_cases/evaluation/llm_math.html