Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -69,7 +69,10 @@ if uploaded_file:
|
|
| 69 |
HF_token = "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"
|
| 70 |
embeddings = HuggingFaceInferenceAPIEmbeddings(api_key=HF_token, model_name="BAAI/bge-base-en-v1.5")
|
| 71 |
|
| 72 |
-
#
|
|
|
|
|
|
|
|
|
|
| 73 |
try:
|
| 74 |
vectorstore = Chroma.from_documents(chunked_documents, embeddings)
|
| 75 |
retriever = vectorstore.as_retriever(search_type="mmr", search_kwargs={"k": 3})
|
|
@@ -81,5 +84,7 @@ if uploaded_file:
|
|
| 81 |
response = answer_with_retrieval(query, retriever)
|
| 82 |
st.write("### Response")
|
| 83 |
st.write(response)
|
|
|
|
|
|
|
| 84 |
except Exception as e:
|
| 85 |
st.error(f"Error creating vector store or generating embeddings: {str(e)}")
|
|
|
|
| 69 |
HF_token = "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"
|
| 70 |
embeddings = HuggingFaceInferenceAPIEmbeddings(api_key=HF_token, model_name="BAAI/bge-base-en-v1.5")
|
| 71 |
|
| 72 |
+
# Debug: Check the length of chunked_documents
|
| 73 |
+
st.write(f"Number of document chunks: {len(chunked_documents)}")
|
| 74 |
+
|
| 75 |
+
# Attempt to create vector store
|
| 76 |
try:
|
| 77 |
vectorstore = Chroma.from_documents(chunked_documents, embeddings)
|
| 78 |
retriever = vectorstore.as_retriever(search_type="mmr", search_kwargs={"k": 3})
|
|
|
|
| 84 |
response = answer_with_retrieval(query, retriever)
|
| 85 |
st.write("### Response")
|
| 86 |
st.write(response)
|
| 87 |
+
except IndexError as ie:
|
| 88 |
+
st.error(f"IndexError during vector store creation: {str(ie)}")
|
| 89 |
except Exception as e:
|
| 90 |
st.error(f"Error creating vector store or generating embeddings: {str(e)}")
|