Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -145,7 +145,7 @@ def unpack_zip_file(filename:str,db:LangChainFAISS,progress):
|
|
| 145 |
|
| 146 |
if 'index.faiss' in [item.lower() for item in contents]:
|
| 147 |
db2 = LangChainFAISS.load_local(tmp_directory, embeddings, allow_dangerous_deserialization=True)
|
| 148 |
-
db2.merge_from(db)
|
| 149 |
return db2
|
| 150 |
|
| 151 |
for file in contents:
|
|
@@ -298,7 +298,7 @@ def ask_llm_stream(system, user_input):
|
|
| 298 |
|
| 299 |
stream = client.chat.completions.create(
|
| 300 |
messages=messages,
|
| 301 |
-
model="mixtral-8x7b-32768",
|
| 302 |
temperature=0.5,
|
| 303 |
max_tokens=1024,
|
| 304 |
top_p=1,
|
|
|
|
| 145 |
|
| 146 |
if 'index.faiss' in [item.lower() for item in contents]:
|
| 147 |
db2 = LangChainFAISS.load_local(tmp_directory, embeddings, allow_dangerous_deserialization=True)
|
| 148 |
+
#db2.merge_from(db)
|
| 149 |
return db2
|
| 150 |
|
| 151 |
for file in contents:
|
|
|
|
| 298 |
|
| 299 |
stream = client.chat.completions.create(
|
| 300 |
messages=messages,
|
| 301 |
+
model="llama3-70b-8192",#"mixtral-8x7b-32768",
|
| 302 |
temperature=0.5,
|
| 303 |
max_tokens=1024,
|
| 304 |
top_p=1,
|