Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -6,6 +6,7 @@ import requests
|
|
| 6 |
import asyncio
|
| 7 |
import json
|
| 8 |
import aiohttp
|
|
|
|
| 9 |
from minivectordb.embedding_model import EmbeddingModel
|
| 10 |
from minivectordb.vector_database import VectorDatabase
|
| 11 |
from text_util_en_pt.cleaner import structurize_text, detect_language, Language
|
|
@@ -17,6 +18,9 @@ torch.set_num_threads(2)
|
|
| 17 |
openrouter_key = os.environ.get("OPENROUTER_KEY")
|
| 18 |
model = EmbeddingModel(use_quantized_onnx_model=True)
|
| 19 |
|
|
|
|
|
|
|
|
|
|
| 20 |
def fetch_links(query, max_results=10):
|
| 21 |
return list(search(query, num_results=max_results))
|
| 22 |
|
|
@@ -71,7 +75,8 @@ def generate_search_terms(message, lang):
|
|
| 71 |
if choices:
|
| 72 |
return choices[0].get('message', {}).get('content', 'Default content if key is missing')
|
| 73 |
else:
|
| 74 |
-
|
|
|
|
| 75 |
|
| 76 |
async def predict(message, history):
|
| 77 |
full_response = ""
|
|
|
|
| 6 |
import asyncio
|
| 7 |
import json
|
| 8 |
import aiohttp
|
| 9 |
+
import logging
|
| 10 |
from minivectordb.embedding_model import EmbeddingModel
|
| 11 |
from minivectordb.vector_database import VectorDatabase
|
| 12 |
from text_util_en_pt.cleaner import structurize_text, detect_language, Language
|
|
|
|
| 18 |
openrouter_key = os.environ.get("OPENROUTER_KEY")
|
| 19 |
model = EmbeddingModel(use_quantized_onnx_model=True)
|
| 20 |
|
| 21 |
+
# Configure logging
|
| 22 |
+
logging.basicConfig(level=logging.ERROR, format='%(asctime)s - %(levelname)s - %(message)s')
|
| 23 |
+
|
| 24 |
def fetch_links(query, max_results=10):
|
| 25 |
return list(search(query, num_results=max_results))
|
| 26 |
|
|
|
|
| 75 |
if choices:
|
| 76 |
return choices[0].get('message', {}).get('content', 'Default content if key is missing')
|
| 77 |
else:
|
| 78 |
+
logging.error(f'No choices available in the response: {response_json}')
|
| 79 |
+
return 'No valid search terms generated'
|
| 80 |
|
| 81 |
async def predict(message, history):
|
| 82 |
full_response = ""
|