import os import time import csv import shutil from datetime import datetime import openai import gradio as gr # Embeddings from langchain.embeddings.openai import OpenAIEmbeddings from langchain.vectorstores import Chroma # Chat Q&A from langchain.chat_models import ChatOpenAI from langchain.schema import AIMessage, HumanMessage, SystemMessage # This sets up OpenAI embeddings model embeddings = OpenAIEmbeddings() # Loads database from persisted directory db_directory = "2023_12_04_chroma_db" db = Chroma(persist_directory=db_directory, embedding_function=embeddings) # This is code that retrieves relevant documents based on a similarity search (in this case, it grabs the top 2 relevant documents or chunks) retriever = db.as_retriever(search_type='similarity', search_kwargs={"k":2}) with open('system_prompt.txt', 'r') as file: ORIG_SYSTEM_MESSAGE_PROMPT = file.read() openai.api_key = os.getenv("OPENAI_API_KEY") #chat = ChatOpenAI(model_name="gpt-3.5-turbo",temperature=0) # Faster for experiments chat = ChatOpenAI(model_name="gpt-3.5-turbo",temperature=0) # Make sure we don't exceed estimation of token limit: TOKEN_LIMIT = 4096 # GPT-3.5 Turbo token limit BUFFER = 100 # Extra tokens to consider for incoming messages PERSISTENT_LOG_PATH = "/data/downvoted_responses.csv" # File in which to log downvoted responses LOCAL_LOG_PATH = "./data/downvoted_responses.csv" def estimate_tokens(texts): return sum([len(t.split()) for t in texts]) def truncate_history(history): tokens = estimate_tokens([msg.content for msg in history]) while tokens + BUFFER > TOKEN_LIMIT and len(history) > 3: history = history[0:1] + history[3:] tokens = estimate_tokens([msg.content for msg in history]) return history # Here is the langchain def predict(history, input): context = retriever.get_relevant_documents(input) print(context) #For debugging history_langchain_format = [] history_langchain_format.append(SystemMessage(content=f"{ORIG_SYSTEM_MESSAGE_PROMPT}")) for human, ai in history: history_langchain_format.append(HumanMessage(content=human)) history_langchain_format.append(AIMessage(content=ai)) history_langchain_format.append(HumanMessage(content=input)) history_langchain_format.append(SystemMessage(content=f"If you need to answer a question based on the previous message, here is some info: {context}")) # Truncate if history is too long history_langchain_format = truncate_history(history_langchain_format) gpt_response = chat(history_langchain_format) # Extract pairs of HumanMessage and AIMessage pairs = [] for i in range(len(history_langchain_format)): if isinstance(history_langchain_format[i], HumanMessage) and (i+1 < len(history_langchain_format)) and isinstance(history_langchain_format[i+1], AIMessage): pairs.append((history_langchain_format[i].content, history_langchain_format[i+1].content)) # Add the new AI response to the pairs for subsequent interactions pairs.append((input, gpt_response.content)) return pairs # Function to handle user message (this clears the interface) def user(user_message, chatbot_history): return "", chatbot_history + [[user_message, ""]] # Function to handle AI's response def bot(chatbot_history): user_message = chatbot_history[-1][0] #This line is because we cleared the user_message previously in the user function above # Call the predict function to get the AI's response pairs = predict(chatbot_history, user_message) _, ai_response = pairs[-1] # Get the latest response response_in_progress = "" for character in ai_response: response_in_progress += character chatbot_history[-1][1] = response_in_progress time.sleep(0.05) yield chatbot_history def log_to_csv(question, answer): """Append a line to a CSV. Create a new file if needed.""" now = datetime.today().strftime("%Y%m%d_%H:%M:%S") if not os.path.isfile(PERSISTENT_LOG_PATH): # Add the column names to the CSV with open(PERSISTENT_LOG_PATH, "w+") as csv_file: writer = csv.writer(csv_file) writer.writerow(["datetime", "user_question", "bot_response"]) # Write the disliked message to the CSV with open(PERSISTENT_LOG_PATH, "a") as csv_file: writer = csv.writer(csv_file) writer.writerow([now, question, answer]) # Copy file from persistent storage to local repo shutil.copyfile(PERSISTENT_LOG_PATH, LOCAL_LOG_PATH) def get_voted_qa_pair(history, voted_answer): """Return the question-answer pair from the chat history, given a particular bot answer. Note: This is required because the 'vote' event handler only has access to the answer that was liked/disliked. """ for question, answer in history: if answer == voted_answer: return question, answer def vote(data: gr.LikeData, history): """This is a function to do something with the voted information""" print(history) if data.liked: print("You upvoted this response: " + data.value) else: print("You downvoted this response: " + data.value) # Find Q/A pair that was disliked question, answer = get_voted_qa_pair(history, data.value) log_to_csv(question, answer) # The Gradio App interface with gr.Blocks() as demo: gr.Markdown("""