Update app.py
Browse files
app.py
CHANGED
|
@@ -285,13 +285,10 @@ def demo():
|
|
| 285 |
collection_name = gr.State()
|
| 286 |
|
| 287 |
gr.Markdown(
|
| 288 |
-
"""<center><h2>
|
| 289 |
-
<h3>
|
| 290 |
gr.Markdown(
|
| 291 |
-
"""<b>Note:</b> This AI assistant, using Langchain and open-source LLMs, performs retrieval-augmented generation (RAG) from your
|
| 292 |
-
The user interface explicitely shows multiple steps to help understand the RAG workflow.
|
| 293 |
-
This chatbot takes past questions into account when generating answers (via conversational memory), and includes document references for clarity purposes.<br>
|
| 294 |
-
<br><b>Warning:</b> This space uses the free CPU Basic hardware from Hugging Face. Some steps and LLM models used below (free inference endpoints) can take some time to generate a reply.
|
| 295 |
""")
|
| 296 |
|
| 297 |
with gr.Tab("Step 1 - Upload PDF"):
|
|
|
|
| 285 |
collection_name = gr.State()
|
| 286 |
|
| 287 |
gr.Markdown(
|
| 288 |
+
"""<center><h2>DocAI</center></h2>
|
| 289 |
+
<h3>Smart Document Insights</h3>""")
|
| 290 |
gr.Markdown(
|
| 291 |
+
"""<b>Note:</b> This AI assistant, using Langchain and open-source LLMs, performs retrieval-augmented generation (RAG) from your documents. \
|
|
|
|
|
|
|
|
|
|
| 292 |
""")
|
| 293 |
|
| 294 |
with gr.Tab("Step 1 - Upload PDF"):
|