Update app.py
Browse files
app.py
CHANGED
|
@@ -208,16 +208,16 @@ def conversation(qa_chain, message, history):
|
|
| 208 |
response_source2 = response_sources[1].page_content.strip()
|
| 209 |
response_source3 = response_sources[2].page_content.strip()
|
| 210 |
# Langchain sources are zero-based
|
| 211 |
-
response_source1_page = response_sources[0].metadata["page"] + 1
|
| 212 |
-
response_source2_page = response_sources[1].metadata["page"] + 1
|
| 213 |
-
response_source3_page = response_sources[2].metadata["page"] + 1
|
| 214 |
# print ('chat response: ', response_answer)
|
| 215 |
# print('DB source', response_sources)
|
| 216 |
|
| 217 |
# Append user message and response to chat history
|
| 218 |
new_history = history + [(message, response_answer)]
|
| 219 |
# return gr.update(value=""), new_history, response_sources[0], response_sources[1]
|
| 220 |
-
return qa_chain, gr.update(value=""), new_history, response_source1, response_source1_page, response_source2, response_source2_page, response_source3, response_source3_page
|
| 221 |
|
| 222 |
|
| 223 |
def upload_file(file_obj):
|
|
@@ -294,7 +294,9 @@ def demo():
|
|
| 294 |
chatbot.change(preprocess = english_to_indian, postprocess = indian_to_english)
|
| 295 |
|
| 296 |
with gr.Row():
|
| 297 |
-
msg = gr.Textbox(placeholder="Type message", container=True)
|
|
|
|
|
|
|
| 298 |
with gr.Accordion("References", open=False):
|
| 299 |
with gr.Row():
|
| 300 |
doc_source1 = gr.Textbox(label="Reference 1", lines=2, container=True, scale=20)
|
|
@@ -305,7 +307,7 @@ def demo():
|
|
| 305 |
with gr.Row():
|
| 306 |
doc_source3 = gr.Textbox(label="Reference 3", lines=2, container=True, scale=20)
|
| 307 |
source3_page = gr.Number(label="Page", scale=1)
|
| 308 |
-
|
| 309 |
with gr.Row():
|
| 310 |
submit_btn = gr.Button("Submit")
|
| 311 |
clear_btn = gr.ClearButton([msg, chatbot])
|
|
@@ -319,21 +321,29 @@ def demo():
|
|
| 319 |
inputs=[slider_temperature, slider_maxtokens, slider_topk, vector_db], \
|
| 320 |
outputs=[qa_chain, llm_progress]).then(lambda:[None,"",0,"",0,"",0], \
|
| 321 |
inputs=None, \
|
| 322 |
-
outputs=[chatbot
|
|
|
|
|
|
|
| 323 |
queue=False)
|
| 324 |
|
| 325 |
# Chatbot events
|
| 326 |
msg.submit(conversation, \
|
| 327 |
inputs=[qa_chain, msg, chatbot], \
|
| 328 |
-
outputs=[qa_chain, msg, chatbot
|
|
|
|
|
|
|
| 329 |
queue=False)
|
| 330 |
submit_btn.click(conversation, \
|
| 331 |
inputs=[qa_chain, msg, chatbot], \
|
| 332 |
-
outputs=[qa_chain, msg, chatbot
|
|
|
|
|
|
|
| 333 |
queue=False)
|
| 334 |
clear_btn.click(lambda:[None,"",0,"",0,"",0], \
|
| 335 |
inputs=None, \
|
| 336 |
-
outputs=[chatbot
|
|
|
|
|
|
|
| 337 |
queue=False)
|
| 338 |
demo.queue().launch(debug=True)
|
| 339 |
|
|
|
|
| 208 |
response_source2 = response_sources[1].page_content.strip()
|
| 209 |
response_source3 = response_sources[2].page_content.strip()
|
| 210 |
# Langchain sources are zero-based
|
| 211 |
+
#response_source1_page = response_sources[0].metadata["page"] + 1
|
| 212 |
+
#response_source2_page = response_sources[1].metadata["page"] + 1
|
| 213 |
+
#response_source3_page = response_sources[2].metadata["page"] + 1
|
| 214 |
# print ('chat response: ', response_answer)
|
| 215 |
# print('DB source', response_sources)
|
| 216 |
|
| 217 |
# Append user message and response to chat history
|
| 218 |
new_history = history + [(message, response_answer)]
|
| 219 |
# return gr.update(value=""), new_history, response_sources[0], response_sources[1]
|
| 220 |
+
return qa_chain, gr.update(value=""), new_history#, response_source1, response_source1_page, response_source2, response_source2_page, response_source3, response_source3_page
|
| 221 |
|
| 222 |
|
| 223 |
def upload_file(file_obj):
|
|
|
|
| 294 |
chatbot.change(preprocess = english_to_indian, postprocess = indian_to_english)
|
| 295 |
|
| 296 |
with gr.Row():
|
| 297 |
+
msg = gr.Textbox(placeholder="Type message", container=True)
|
| 298 |
+
|
| 299 |
+
'''
|
| 300 |
with gr.Accordion("References", open=False):
|
| 301 |
with gr.Row():
|
| 302 |
doc_source1 = gr.Textbox(label="Reference 1", lines=2, container=True, scale=20)
|
|
|
|
| 307 |
with gr.Row():
|
| 308 |
doc_source3 = gr.Textbox(label="Reference 3", lines=2, container=True, scale=20)
|
| 309 |
source3_page = gr.Number(label="Page", scale=1)
|
| 310 |
+
'''
|
| 311 |
with gr.Row():
|
| 312 |
submit_btn = gr.Button("Submit")
|
| 313 |
clear_btn = gr.ClearButton([msg, chatbot])
|
|
|
|
| 321 |
inputs=[slider_temperature, slider_maxtokens, slider_topk, vector_db], \
|
| 322 |
outputs=[qa_chain, llm_progress]).then(lambda:[None,"",0,"",0,"",0], \
|
| 323 |
inputs=None, \
|
| 324 |
+
outputs=[chatbot
|
| 325 |
+
#, doc_source1, source1_page, doc_source2, source2_page, doc_source3, source3_page
|
| 326 |
+
], \
|
| 327 |
queue=False)
|
| 328 |
|
| 329 |
# Chatbot events
|
| 330 |
msg.submit(conversation, \
|
| 331 |
inputs=[qa_chain, msg, chatbot], \
|
| 332 |
+
outputs=[qa_chain, msg, chatbot
|
| 333 |
+
#, doc_source1, source1_page, doc_source2, source2_page, doc_source3, source3_page
|
| 334 |
+
], \
|
| 335 |
queue=False)
|
| 336 |
submit_btn.click(conversation, \
|
| 337 |
inputs=[qa_chain, msg, chatbot], \
|
| 338 |
+
outputs=[qa_chain, msg, chatbot
|
| 339 |
+
#, doc_source1, source1_page, doc_source2, source2_page, doc_source3, source3_page
|
| 340 |
+
], \
|
| 341 |
queue=False)
|
| 342 |
clear_btn.click(lambda:[None,"",0,"",0,"",0], \
|
| 343 |
inputs=None, \
|
| 344 |
+
outputs=[chatbot
|
| 345 |
+
#, doc_source1, source1_page, doc_source2, source2_page, doc_source3, source3_page
|
| 346 |
+
], \
|
| 347 |
queue=False)
|
| 348 |
demo.queue().launch(debug=True)
|
| 349 |
|