Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -907,7 +907,8 @@ REMEMBER: One tool per turn. No reasoning without tools. Exact answer format.
|
|
| 907 |
# max_new_tokens=2048, # Increased token limit for potentially longer reasoning/tool use
|
| 908 |
# temperature=0.01, # Keep temperature low for factual tasks
|
| 909 |
#Use tool_choice="any" to FORCE tool usage
|
| 910 |
-
"""
|
|
|
|
| 911 |
try:
|
| 912 |
self.llm_with_tools = ChatGroq(
|
| 913 |
temperature=0,
|
|
@@ -920,11 +921,12 @@ REMEMBER: One tool per turn. No reasoning without tools. Exact answer format.
|
|
| 920 |
|
| 921 |
except Exception as e:
|
| 922 |
print(f"❌ Error initializing HuggingFace: {e}")
|
| 923 |
-
raise
|
| 924 |
-
|
|
|
|
| 925 |
try:
|
| 926 |
self.llm_with_tools = ChatGoogleGenerativeAI(
|
| 927 |
-
model="gemini-2.
|
| 928 |
google_api_key=GOOGLE_API_KEY,
|
| 929 |
temperature=0,
|
| 930 |
max_output_tokens=8192,
|
|
|
|
| 907 |
# max_new_tokens=2048, # Increased token limit for potentially longer reasoning/tool use
|
| 908 |
# temperature=0.01, # Keep temperature low for factual tasks
|
| 909 |
#Use tool_choice="any" to FORCE tool usage
|
| 910 |
+
"""
|
| 911 |
+
print("Initializing Groq LLM...")
|
| 912 |
try:
|
| 913 |
self.llm_with_tools = ChatGroq(
|
| 914 |
temperature=0,
|
|
|
|
| 921 |
|
| 922 |
except Exception as e:
|
| 923 |
print(f"❌ Error initializing HuggingFace: {e}")
|
| 924 |
+
raise
|
| 925 |
+
"""
|
| 926 |
+
print("Initializing Google Gemini LLM...")
|
| 927 |
try:
|
| 928 |
self.llm_with_tools = ChatGoogleGenerativeAI(
|
| 929 |
+
model="gemini-2.5-flash", # Latest model
|
| 930 |
google_api_key=GOOGLE_API_KEY,
|
| 931 |
temperature=0,
|
| 932 |
max_output_tokens=8192,
|