Update app.py
Browse files
app.py
CHANGED
|
@@ -64,9 +64,14 @@ def setup_multiple_chains():
|
|
| 64 |
)
|
| 65 |
|
| 66 |
conversation_memory = ConversationBufferMemory(
|
| 67 |
-
memory_key="chat_history",
|
|
|
|
|
|
|
|
|
|
|
|
|
| 68 |
)
|
| 69 |
|
|
|
|
| 70 |
llm_chain = LLMChain(
|
| 71 |
llm=llm,
|
| 72 |
prompt=daysoff_assistant_prompt,
|
|
@@ -118,9 +123,19 @@ async def handle_message(message: cl.Message):
|
|
| 118 |
except requests.exceptions.RequestException as e:
|
| 119 |
await cl.Message(content=f"Request failed: {str(e)}").send()
|
| 120 |
else:
|
| 121 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 122 |
|
| 123 |
|
| 124 |
-
response_key = "output" if "output" in response else "text"
|
| 125 |
-
await cl.Message(response.get(response_key, "")).send()
|
| 126 |
-
return message.content
|
|
|
|
| 64 |
)
|
| 65 |
|
| 66 |
conversation_memory = ConversationBufferMemory(
|
| 67 |
+
memory_key="chat_history",
|
| 68 |
+
input_key="question", # ?
|
| 69 |
+
output_key="text", # ?
|
| 70 |
+
max_len=30,
|
| 71 |
+
return_messages=True
|
| 72 |
)
|
| 73 |
|
| 74 |
+
|
| 75 |
llm_chain = LLMChain(
|
| 76 |
llm=llm,
|
| 77 |
prompt=daysoff_assistant_prompt,
|
|
|
|
| 123 |
except requests.exceptions.RequestException as e:
|
| 124 |
await cl.Message(content=f"Request failed: {str(e)}").send()
|
| 125 |
else:
|
| 126 |
+
try:
|
| 127 |
+
response = await llm_chain.ainvoke({
|
| 128 |
+
"question": user_message,
|
| 129 |
+
"chat_history": ""
|
| 130 |
+
}, callbacks=[cl.AsyncLangchainCallbackHandler()])
|
| 131 |
+
|
| 132 |
+
await cl.Message(content=response["text"]).send()
|
| 133 |
+
|
| 134 |
+
except Exception as e:
|
| 135 |
+
await cl.Message(content=f"Error: {str(e)}").send()
|
| 136 |
+
|
| 137 |
|
| 138 |
|
| 139 |
+
#response_key = "output" if "output" in response else "text"
|
| 140 |
+
#await cl.Message(response.get(response_key, "")).send()
|
| 141 |
+
#return message.content
|