Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -28,6 +28,9 @@ client = InferenceClient("mistralai/Mistral-7B-Instruct-v0.2")
|
|
| 28 |
# --- RAG pipeline ---
|
| 29 |
def generate_answer(messages):
|
| 30 |
try:
|
|
|
|
|
|
|
|
|
|
| 31 |
user_query = messages[-1]["content"]
|
| 32 |
query_embedding = embedding_model.encode([user_query], convert_to_numpy=True)
|
| 33 |
faiss.normalize_L2(query_embedding)
|
|
|
|
| 28 |
# --- RAG pipeline ---
|
| 29 |
def generate_answer(messages):
|
| 30 |
try:
|
| 31 |
+
sys.stderr.write("=== generate_answer called ===\n")
|
| 32 |
+
sys.stderr.flush()
|
| 33 |
+
|
| 34 |
user_query = messages[-1]["content"]
|
| 35 |
query_embedding = embedding_model.encode([user_query], convert_to_numpy=True)
|
| 36 |
faiss.normalize_L2(query_embedding)
|