Spaces:
Sleeping
Sleeping
app18
Browse files
app.py
CHANGED
|
@@ -111,7 +111,7 @@ qa_chain = ConversationalRetrievalChain.from_llm(
|
|
| 111 |
get_chat_history = lambda h : h
|
| 112 |
)
|
| 113 |
|
| 114 |
-
def chat_interface(question):
|
| 115 |
|
| 116 |
result = qa_chain.invoke({"question": question})
|
| 117 |
|
|
|
|
| 111 |
get_chat_history = lambda h : h
|
| 112 |
)
|
| 113 |
|
| 114 |
+
def chat_interface(question,history):
|
| 115 |
|
| 116 |
result = qa_chain.invoke({"question": question})
|
| 117 |
|