Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -773,13 +773,14 @@ Your goal: Provide the EXACT answer in the EXACT format requested.
|
|
| 773 |
print("Initializing Google Gemini LLM...")
|
| 774 |
try:
|
| 775 |
# --- Initialize ONE Powerful LLM for all tasks ---
|
| 776 |
-
self.
|
| 777 |
-
model="gemini-
|
| 778 |
google_api_key=GEMINI_API_KEY,
|
| 779 |
temperature=0,
|
| 780 |
-
convert_system_message_to_human=True
|
| 781 |
-
|
| 782 |
-
|
|
|
|
| 783 |
|
| 784 |
except Exception as e:
|
| 785 |
print(f"β Error initializing Gemini: {e}")
|
|
|
|
| 773 |
print("Initializing Google Gemini LLM...")
|
| 774 |
try:
|
| 775 |
# --- Initialize ONE Powerful LLM for all tasks ---
|
| 776 |
+
self.planner_llm = ChatGoogleGenerativeAI(
|
| 777 |
+
model="gemini-2.0-flash-exp", # β Same model, different config
|
| 778 |
google_api_key=GEMINI_API_KEY,
|
| 779 |
temperature=0,
|
| 780 |
+
convert_system_message_to_human=True,
|
| 781 |
+
max_output_tokens=1024 # Planner needs less tokens
|
| 782 |
+
)
|
| 783 |
+
print("β
Planner LLM (Gemini 2.0 Flash) initialized.")
|
| 784 |
|
| 785 |
except Exception as e:
|
| 786 |
print(f"β Error initializing Gemini: {e}")
|