Spaces:
Sleeping
Sleeping
scrollable UI attempt
Browse files
app.py
CHANGED
|
@@ -112,6 +112,8 @@ def generate_concise_response(prompt, context):
|
|
| 112 |
|
| 113 |
# Main function to execute the pipeline
|
| 114 |
def chatbot(message, history):
|
|
|
|
|
|
|
| 115 |
lecture_notes = fetch_lecture_notes()
|
| 116 |
model_architectures = fetch_model_architectures()
|
| 117 |
|
|
@@ -141,12 +143,15 @@ def chatbot(message, history):
|
|
| 141 |
|
| 142 |
# Generate a concise and relevant summary using Gemini
|
| 143 |
prompt = "Summarize the user queries so far"
|
| 144 |
-
user_queries_summary = " ".join(message)
|
| 145 |
concise_response = generate_concise_response(prompt, user_queries_summary)
|
| 146 |
print("Concise Response:")
|
| 147 |
print(concise_response)
|
| 148 |
|
| 149 |
-
|
|
|
|
|
|
|
|
|
|
| 150 |
|
| 151 |
iface = gr.ChatInterface(
|
| 152 |
chatbot,
|
|
|
|
| 112 |
|
| 113 |
# Main function to execute the pipeline
|
| 114 |
def chatbot(message, history):
|
| 115 |
+
global conversation_history
|
| 116 |
+
|
| 117 |
lecture_notes = fetch_lecture_notes()
|
| 118 |
model_architectures = fetch_model_architectures()
|
| 119 |
|
|
|
|
| 143 |
|
| 144 |
# Generate a concise and relevant summary using Gemini
|
| 145 |
prompt = "Summarize the user queries so far"
|
| 146 |
+
user_queries_summary = " ".join([message])
|
| 147 |
concise_response = generate_concise_response(prompt, user_queries_summary)
|
| 148 |
print("Concise Response:")
|
| 149 |
print(concise_response)
|
| 150 |
|
| 151 |
+
# Update conversation history with the new user message and system response
|
| 152 |
+
history.append((message, total_text))
|
| 153 |
+
|
| 154 |
+
return history
|
| 155 |
|
| 156 |
iface = gr.ChatInterface(
|
| 157 |
chatbot,
|