Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -122,8 +122,21 @@ def search_glossary(query, roleplaying_glossary):
|
|
| 122 |
|
| 123 |
if not found:
|
| 124 |
st.write("No results found.")
|
| 125 |
-
|
| 126 |
-
st.write('## Processing query against
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 127 |
|
| 128 |
|
| 129 |
|
|
|
|
| 122 |
|
| 123 |
if not found:
|
| 124 |
st.write("No results found.")
|
| 125 |
+
|
| 126 |
+
st.write('## Processing query against GPT and Llama:')
|
| 127 |
+
# ------------------------------------------------------------------------------------------------
|
| 128 |
+
st.write('Reasoning with your inputs using GPT...')
|
| 129 |
+
response = chat_with_model(query)
|
| 130 |
+
st.write('Response:')
|
| 131 |
+
st.write(response)
|
| 132 |
+
filename = generate_filename(response, "txt")
|
| 133 |
+
create_file(filename, query, query + ' --- ' + response, should_save)
|
| 134 |
+
|
| 135 |
+
st.write('Reasoning with your inputs using Llama...')
|
| 136 |
+
response = StreamLLMChatResponse(query)
|
| 137 |
+
filename_txt = generate_filename(query, "md")
|
| 138 |
+
create_file(filename_txt, query, query + ' --- ' + response, should_save)
|
| 139 |
+
# ------------------------------------------------------------------------------------------------
|
| 140 |
|
| 141 |
|
| 142 |
|