Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -32,7 +32,7 @@ def get_vectorstore(text_chunks):
|
|
| 32 |
model = "BAAI/bge-base-en-v1.5"
|
| 33 |
encode_kwargs = {
|
| 34 |
"normalize_embeddings": True
|
| 35 |
-
}
|
| 36 |
embeddings = HuggingFaceBgeEmbeddings(
|
| 37 |
model_name=model, encode_kwargs=encode_kwargs, model_kwargs={"device": "cpu"}
|
| 38 |
)
|
|
@@ -45,7 +45,6 @@ def get_conversation_chain(vectorstore):
|
|
| 45 |
repo_id="mistralai/Mixtral-8x7B-Instruct-v0.1",
|
| 46 |
model_kwargs={"temperature": 0.5, "max_length": 1048},
|
| 47 |
)
|
| 48 |
-
# llm = ChatOpenAI(temperature=0, model="gpt-3.5-turbo-0613")
|
| 49 |
|
| 50 |
memory = ConversationBufferMemory(memory_key="chat_history", return_messages=True)
|
| 51 |
conversation_chain = ConversationalRetrievalChain.from_llm(
|
|
|
|
| 32 |
model = "BAAI/bge-base-en-v1.5"
|
| 33 |
encode_kwargs = {
|
| 34 |
"normalize_embeddings": True
|
| 35 |
+
}
|
| 36 |
embeddings = HuggingFaceBgeEmbeddings(
|
| 37 |
model_name=model, encode_kwargs=encode_kwargs, model_kwargs={"device": "cpu"}
|
| 38 |
)
|
|
|
|
| 45 |
repo_id="mistralai/Mixtral-8x7B-Instruct-v0.1",
|
| 46 |
model_kwargs={"temperature": 0.5, "max_length": 1048},
|
| 47 |
)
|
|
|
|
| 48 |
|
| 49 |
memory = ConversationBufferMemory(memory_key="chat_history", return_messages=True)
|
| 50 |
conversation_chain = ConversationalRetrievalChain.from_llm(
|