Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -1,93 +1,40 @@
|
|
| 1 |
import streamlit as st
|
|
|
|
| 2 |
from langchain_core.messages import AIMessage, HumanMessage
|
| 3 |
-
from langchain_openai import ChatOpenAI
|
| 4 |
-
|
| 5 |
-
from langchain_core.output_parsers import StrOutputParser
|
| 6 |
-
from langchain_core.prompts import ChatPromptTemplate
|
| 7 |
|
| 8 |
-
|
| 9 |
-
from langchain_openai import AzureChatOpenAI
|
| 10 |
-
from langchain.schema import HumanMessage, SystemMessage
|
| 11 |
-
from langchain_core.prompts.chat import SystemMessagePromptTemplate, HumanMessagePromptTemplate, ChatPromptTemplate
|
| 12 |
-
from azure_openai import qt
|
| 13 |
from retriver import search_and_reconstruct
|
| 14 |
-
# Initialize an instance of AzureOpenAI using the specified settings
|
| 15 |
-
import pandas as pd
|
| 16 |
-
|
| 17 |
-
|
| 18 |
-
# LLM Langchain Definition
|
| 19 |
-
OPENAI_API_KEY = st.secrets['azure_api_key']
|
| 20 |
-
OPENAI_API_TYPE = "azure"
|
| 21 |
-
OPENAI_API_BASE = "https://davidfearn-gpt4.openai.azure.com"
|
| 22 |
-
# OPENAI_API_VERSION = "2024-02-01"
|
| 23 |
-
OPENAI_API_VERSION = "2024-08-01-preview"
|
| 24 |
-
# OPENAI_MODEL = "gpt4-turbo-1106"
|
| 25 |
-
OPENAI_MODEL = "gpt-4o"
|
| 26 |
-
# Initialize an instance of AzureOpenAI using the specified settings
|
| 27 |
|
|
|
|
|
|
|
|
|
|
| 28 |
|
| 29 |
-
def get_response(chat_history, qte, knowledge, temp1, temp2, tokens1, tokens2, persona2SystemMessage, persona2UserMessage):
|
| 30 |
-
|
| 31 |
-
llm = AzureChatOpenAI(
|
| 32 |
-
openai_api_version=OPENAI_API_VERSION,
|
| 33 |
-
openai_api_key=OPENAI_API_KEY,
|
| 34 |
-
azure_endpoint=OPENAI_API_BASE,
|
| 35 |
-
openai_api_type=OPENAI_API_TYPE,
|
| 36 |
-
deployment_name=OPENAI_MODEL,
|
| 37 |
-
temperature=temp2,
|
| 38 |
-
max_tokens=tokens2
|
| 39 |
-
# Name of the deployment for identification
|
| 40 |
-
)
|
| 41 |
-
|
| 42 |
-
|
| 43 |
-
system_message_template = SystemMessagePromptTemplate.from_template(persona2SystemMessage)
|
| 44 |
-
# human_message_template = HumanMessagePromptTemplate.from_template(persona2UserMessage)
|
| 45 |
-
|
| 46 |
-
# Create a chat prompt template combining system and human messages
|
| 47 |
-
# prompt = ChatPromptTemplate.from_messages([system_message_template, human_message_template])
|
| 48 |
-
prompt = ChatPromptTemplate.from_messages([system_message_template, persona2UserMessage])
|
| 49 |
-
|
| 50 |
-
chain = prompt | llm | StrOutputParser()
|
| 51 |
-
|
| 52 |
-
return chain.stream({
|
| 53 |
-
"query": chat_history,
|
| 54 |
-
"knowledge": knowledge
|
| 55 |
-
})
|
| 56 |
|
| 57 |
-
placeHolderPersona1system = "place holder"
|
| 58 |
-
placeHolderPersona2user = "Query: {query}, Knowledge: {knowledge}"
|
| 59 |
# app config
|
| 60 |
st.set_page_config(page_title="Reg Intel Chatbot", page_icon="🤖")
|
| 61 |
st.title("Reg Intel Toolbox :toolbox:")
|
| 62 |
|
| 63 |
# Sidebar for inputting personas
|
| 64 |
st.sidebar.title("RAG System Designer")
|
| 65 |
-
# st.sidebar.subheader("Welcome Message")
|
| 66 |
-
# welcomeMessage = st.sidebar.text_area("Define Intake Persona", value=welcomeMessage, height=300)
|
| 67 |
st.sidebar.header("Query Designer Config")
|
| 68 |
-
|
| 69 |
-
persona1SystemMessage = st.sidebar.text_area("Query Designer System Message", value=placeHolderPersona1system, height=300)
|
| 70 |
temp1 = st.sidebar.slider("Temperature", min_value=0.0, max_value=1.0, step=0.1, value=0.6, key='persona1_temp')
|
| 71 |
tokens1 = st.sidebar.slider("Tokens", min_value=0, max_value=4000, step=100, value=500, key='persona1_tokens')
|
| 72 |
st.sidebar.subheader("Number of Search Results")
|
| 73 |
k = st.sidebar.slider("Returned Docs", min_value=1, max_value=10, step=1, value=3, key='k')
|
| 74 |
-
pagesReturned = st.sidebar.slider("Number of
|
| 75 |
-
|
| 76 |
-
|
| 77 |
st.sidebar.header("Engineered Prompt Config")
|
| 78 |
-
persona2SystemMessage = st.sidebar.text_area("Answer Creation System Message", value=
|
| 79 |
-
persona2UserMessage = st.sidebar.text_area("Answer Creation User Message", value=
|
| 80 |
temp2 = st.sidebar.slider("Temperature", min_value=0.0, max_value=1.0, step=0.1, value=0.6, key='persona2_temp')
|
| 81 |
tokens2 = st.sidebar.slider("Tokens", min_value=0, max_value=4000, step=100, value=500, key='persona2_tokens')
|
| 82 |
|
| 83 |
-
|
| 84 |
# session state
|
| 85 |
if "chat_history" not in st.session_state:
|
| 86 |
st.session_state.chat_history = [
|
| 87 |
AIMessage(content="Hello, I am the GSK Reg Intel Assistant. How can I help you?"),
|
| 88 |
]
|
| 89 |
|
| 90 |
-
|
| 91 |
# conversation
|
| 92 |
for message in st.session_state.chat_history:
|
| 93 |
if isinstance(message, AIMessage):
|
|
@@ -97,7 +44,6 @@ for message in st.session_state.chat_history:
|
|
| 97 |
with st.chat_message("Human"):
|
| 98 |
st.write(message.content)
|
| 99 |
|
| 100 |
-
|
| 101 |
# user input
|
| 102 |
user_query = st.chat_input("Type your message here...")
|
| 103 |
if user_query is not None and user_query != "":
|
|
@@ -106,15 +52,12 @@ if user_query is not None and user_query != "":
|
|
| 106 |
with st.chat_message("Human"):
|
| 107 |
st.markdown(user_query)
|
| 108 |
|
| 109 |
-
|
| 110 |
-
|
| 111 |
-
knowledge = search_and_reconstruct(qte, k, pagesReturned)
|
| 112 |
-
response = st.write_stream(get_response(st.session_state.chat_history, qte, knowledge, temp1, temp2, tokens1, tokens2, persona2SystemMessage, persona2UserMessage))
|
| 113 |
-
|
| 114 |
-
st.session_state.chat_history.append(AIMessage(content=response))
|
| 115 |
st.sidebar.header("QTE and Knowledge Results")
|
| 116 |
st.sidebar.header("QTE")
|
| 117 |
st.sidebar.text(qte)
|
|
|
|
| 118 |
|
| 119 |
if knowledge:
|
| 120 |
|
|
@@ -134,4 +77,8 @@ if user_query is not None and user_query != "":
|
|
| 134 |
st.sidebar.write("### Knowledge Base Results")
|
| 135 |
st.sidebar.dataframe(df) # Adjust height as needed
|
| 136 |
else:
|
| 137 |
-
st.sidebar.write("No relevant knowledge base results found.")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
import streamlit as st
|
| 2 |
+
import pandas as pd
|
| 3 |
from langchain_core.messages import AIMessage, HumanMessage
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
|
| 5 |
+
from azure_openai import qt, get_response
|
|
|
|
|
|
|
|
|
|
|
|
|
| 6 |
from retriver import search_and_reconstruct
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 7 |
|
| 8 |
+
placeHolderPersonaQTE = "{assetGlossary}, {conversationToDate}"
|
| 9 |
+
placeHolderPersonaRAGSystem = "{assetGlossary}"
|
| 10 |
+
placeHolderPersonaRAGUser = "{query}, {knowledge}"
|
| 11 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 12 |
|
|
|
|
|
|
|
| 13 |
# app config
|
| 14 |
st.set_page_config(page_title="Reg Intel Chatbot", page_icon="🤖")
|
| 15 |
st.title("Reg Intel Toolbox :toolbox:")
|
| 16 |
|
| 17 |
# Sidebar for inputting personas
|
| 18 |
st.sidebar.title("RAG System Designer")
|
|
|
|
|
|
|
| 19 |
st.sidebar.header("Query Designer Config")
|
| 20 |
+
persona1SystemMessage = st.sidebar.text_area("Query Designer System Message", value=placeHolderPersonaQTE, height=300)
|
|
|
|
| 21 |
temp1 = st.sidebar.slider("Temperature", min_value=0.0, max_value=1.0, step=0.1, value=0.6, key='persona1_temp')
|
| 22 |
tokens1 = st.sidebar.slider("Tokens", min_value=0, max_value=4000, step=100, value=500, key='persona1_tokens')
|
| 23 |
st.sidebar.subheader("Number of Search Results")
|
| 24 |
k = st.sidebar.slider("Returned Docs", min_value=1, max_value=10, step=1, value=3, key='k')
|
| 25 |
+
pagesReturned = st.sidebar.slider("Number of Returned Pages", min_value=1, max_value=10, step=1, value=1, key='pagesReturned')
|
|
|
|
|
|
|
| 26 |
st.sidebar.header("Engineered Prompt Config")
|
| 27 |
+
persona2SystemMessage = st.sidebar.text_area("Answer Creation System Message", value=placeHolderPersonaRAGSystem, height=300)
|
| 28 |
+
persona2UserMessage = st.sidebar.text_area("Answer Creation User Message", value=placeHolderPersonaRAGUser, height=300)
|
| 29 |
temp2 = st.sidebar.slider("Temperature", min_value=0.0, max_value=1.0, step=0.1, value=0.6, key='persona2_temp')
|
| 30 |
tokens2 = st.sidebar.slider("Tokens", min_value=0, max_value=4000, step=100, value=500, key='persona2_tokens')
|
| 31 |
|
|
|
|
| 32 |
# session state
|
| 33 |
if "chat_history" not in st.session_state:
|
| 34 |
st.session_state.chat_history = [
|
| 35 |
AIMessage(content="Hello, I am the GSK Reg Intel Assistant. How can I help you?"),
|
| 36 |
]
|
| 37 |
|
|
|
|
| 38 |
# conversation
|
| 39 |
for message in st.session_state.chat_history:
|
| 40 |
if isinstance(message, AIMessage):
|
|
|
|
| 44 |
with st.chat_message("Human"):
|
| 45 |
st.write(message.content)
|
| 46 |
|
|
|
|
| 47 |
# user input
|
| 48 |
user_query = st.chat_input("Type your message here...")
|
| 49 |
if user_query is not None and user_query != "":
|
|
|
|
| 52 |
with st.chat_message("Human"):
|
| 53 |
st.markdown(user_query)
|
| 54 |
|
| 55 |
+
|
| 56 |
+
qte = qt(persona1SystemMessage, st.session_state.chat_history, temp1, tokens1, "GSKGlossary")
|
|
|
|
|
|
|
|
|
|
|
|
|
| 57 |
st.sidebar.header("QTE and Knowledge Results")
|
| 58 |
st.sidebar.header("QTE")
|
| 59 |
st.sidebar.text(qte)
|
| 60 |
+
knowledge = search_and_reconstruct(qte, k, pagesReturned)
|
| 61 |
|
| 62 |
if knowledge:
|
| 63 |
|
|
|
|
| 77 |
st.sidebar.write("### Knowledge Base Results")
|
| 78 |
st.sidebar.dataframe(df) # Adjust height as needed
|
| 79 |
else:
|
| 80 |
+
st.sidebar.write("No relevant knowledge base results found.")
|
| 81 |
+
|
| 82 |
+
with st.chat_message("AI"):
|
| 83 |
+
response = st.write_stream(get_response(st.session_state.chat_history, qte, knowledge, temp1, temp2, tokens1, tokens2, persona2SystemMessage, persona2UserMessage, "GSKGlossary"))
|
| 84 |
+
st.session_state.chat_history.append(AIMessage(content=response))
|