Commit ·
c36041c
1
Parent(s): f31fcbe
Update app.py
Browse files
app.py
CHANGED
|
@@ -92,9 +92,9 @@ def predict(message, model_type):
|
|
| 92 |
# conversational retrieval agent component construction - memory, prompt template, agent, agent executor
|
| 93 |
# specifying LLM to use
|
| 94 |
if (model_type==1):
|
| 95 |
-
llm = ChatOpenAI(temperature = 0.1, model_name="gpt-3.5-turbo-16k", streaming=True, callbacks=[QueueCallback(q)])
|
| 96 |
else:
|
| 97 |
-
llm = ChatOpenAI(temperature = 0.1, model_name="gpt-4", streaming=True, callbacks=[QueueCallback(q)])
|
| 98 |
|
| 99 |
# This is needed for both the memory and the prompt
|
| 100 |
memory_key = "history"
|
|
|
|
| 92 |
# conversational retrieval agent component construction - memory, prompt template, agent, agent executor
|
| 93 |
# specifying LLM to use
|
| 94 |
if (model_type==1):
|
| 95 |
+
llm = ChatOpenAI(temperature = 0.1, model_name="gpt-3.5-turbo-16k-0613", streaming=True, callbacks=[QueueCallback(q)])
|
| 96 |
else:
|
| 97 |
+
llm = ChatOpenAI(temperature = 0.1, model_name="gpt-4-0613", streaming=True, callbacks=[QueueCallback(q)])
|
| 98 |
|
| 99 |
# This is needed for both the memory and the prompt
|
| 100 |
memory_key = "history"
|