Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -43,7 +43,7 @@ def build_faiss_index(chunks, model):
|
|
| 43 |
for chunk in chunks:
|
| 44 |
input_ids = torch.tensor([chunk])
|
| 45 |
with torch.no_grad():
|
| 46 |
-
embedding = model(input_ids).last_hidden_state.mean(dim=1).numpy()
|
| 47 |
embeddings.append(embedding)
|
| 48 |
embeddings = np.vstack(embeddings)
|
| 49 |
|
|
@@ -79,7 +79,9 @@ if text:
|
|
| 79 |
st.write("Searching for the most relevant chunk...")
|
| 80 |
query_tokens = tokenizer.encode(query, add_special_tokens=False)
|
| 81 |
query_embedding = (
|
| 82 |
-
model(torch.tensor([query_tokens]))
|
|
|
|
|
|
|
| 83 |
)
|
| 84 |
_, indices = index.search(query_embedding, k=1)
|
| 85 |
|
|
@@ -104,4 +106,3 @@ else:
|
|
| 104 |
st.error("Failed to extract content from the document.")
|
| 105 |
|
| 106 |
|
| 107 |
-
|
|
|
|
| 43 |
for chunk in chunks:
|
| 44 |
input_ids = torch.tensor([chunk])
|
| 45 |
with torch.no_grad():
|
| 46 |
+
embedding = model(input_ids).last_hidden_state.mean(dim=1).detach().numpy()
|
| 47 |
embeddings.append(embedding)
|
| 48 |
embeddings = np.vstack(embeddings)
|
| 49 |
|
|
|
|
| 79 |
st.write("Searching for the most relevant chunk...")
|
| 80 |
query_tokens = tokenizer.encode(query, add_special_tokens=False)
|
| 81 |
query_embedding = (
|
| 82 |
+
model(torch.tensor([query_tokens]))
|
| 83 |
+
.last_hidden_state.mean(dim=1)
|
| 84 |
+
.detach().numpy()
|
| 85 |
)
|
| 86 |
_, indices = index.search(query_embedding, k=1)
|
| 87 |
|
|
|
|
| 106 |
st.error("Failed to extract content from the document.")
|
| 107 |
|
| 108 |
|
|
|