adjust parallel search and rerank
Browse files- app/gemini_client.py +5 -0
app/gemini_client.py
CHANGED
|
@@ -9,6 +9,7 @@ from .request_limit_manager import RequestLimitManager
|
|
| 9 |
from .utils import (
|
| 10 |
_safe_truncate
|
| 11 |
)
|
|
|
|
| 12 |
|
| 13 |
class GeminiResponseError(Exception):
|
| 14 |
"""Custom exception for non-retriable Gemini response issues like safety or token limits."""
|
|
@@ -24,6 +25,10 @@ class GeminiResponseError(Exception):
|
|
| 24 |
class GeminiClient:
|
| 25 |
def __init__(self):
|
| 26 |
self.limit_manager = RequestLimitManager("gemini")
|
|
|
|
|
|
|
|
|
|
|
|
|
| 27 |
self._cached_model = None
|
| 28 |
self._cached_key = None
|
| 29 |
self._cached_model_instance = None
|
|
|
|
| 9 |
from .utils import (
|
| 10 |
_safe_truncate
|
| 11 |
)
|
| 12 |
+
from .config import get_settings
|
| 13 |
|
| 14 |
class GeminiResponseError(Exception):
|
| 15 |
"""Custom exception for non-retriable Gemini response issues like safety or token limits."""
|
|
|
|
| 25 |
class GeminiClient:
|
| 26 |
def __init__(self):
|
| 27 |
self.limit_manager = RequestLimitManager("gemini")
|
| 28 |
+
settings = get_settings()
|
| 29 |
+
num_keys = len(settings.gemini_api_keys.split(',')) if settings.gemini_api_keys else 0
|
| 30 |
+
num_models = len(settings.gemini_models.split(',')) if settings.gemini_models else 0
|
| 31 |
+
logger.info(f"[GEMINI_INIT] Limiter is considering {num_keys} API keys and {num_models} models.")
|
| 32 |
self._cached_model = None
|
| 33 |
self._cached_key = None
|
| 34 |
self._cached_model_instance = None
|