Spaces:
Sleeping
Sleeping
Fix preset override issue - always use current UI values for execution
Browse files- app_with_mcp.py +20 -25
app_with_mcp.py
CHANGED
|
@@ -191,38 +191,33 @@ async def minion_respond_async(query: str, preset_model: str, api_type: str,
|
|
| 191 |
use_tools: bool):
|
| 192 |
"""Respond to query using specified configuration with optional MCP tools"""
|
| 193 |
|
| 194 |
-
#
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 195 |
if preset_model != "Custom":
|
| 196 |
-
|
| 197 |
-
|
| 198 |
-
|
| 199 |
-
|
| 200 |
-
|
| 201 |
-
|
| 202 |
-
|
| 203 |
-
'model': config_obj.model,
|
| 204 |
-
'temperature': temperature, # Always use UI value for dynamic parameters
|
| 205 |
-
'max_tokens': max_tokens, # Always use UI value for dynamic parameters
|
| 206 |
-
'vision_enabled': config_obj.vision_enabled
|
| 207 |
-
}
|
| 208 |
-
print(f"🔧 Using preset '{preset_model}' with overrides:")
|
| 209 |
-
print(f" - Temperature: {temperature} (preset: {config_obj.temperature})")
|
| 210 |
-
print(f" - Max tokens: {max_tokens} (preset: {config_obj.max_tokens})")
|
| 211 |
-
print(f" - Model: {config_obj.model}")
|
| 212 |
-
print(f" - API Type: {config_obj.api_type}")
|
| 213 |
else:
|
| 214 |
-
# Use completely custom configuration
|
| 215 |
-
llm_config_dict = create_custom_llm_config(
|
| 216 |
-
api_type, api_key, base_url, api_version, model, temperature, max_tokens
|
| 217 |
-
)
|
| 218 |
print(f"🔧 Using custom configuration:")
|
|
|
|
|
|
|
|
|
|
|
|
|
| 219 |
print(f" - Temperature: {temperature}")
|
| 220 |
print(f" - Max tokens: {max_tokens}")
|
| 221 |
-
print(f" - Model: {model}")
|
| 222 |
-
print(f" - API Type: {api_type}")
|
| 223 |
|
| 224 |
-
# Always rebuild brain with current configuration
|
| 225 |
print(f"🧠 Building brain with final config:")
|
|
|
|
|
|
|
| 226 |
print(f" - Final temperature: {llm_config_dict['temperature']}")
|
| 227 |
print(f" - Final max_tokens: {llm_config_dict['max_tokens']}")
|
| 228 |
brain = build_brain_with_config(llm_config_dict)
|
|
|
|
| 191 |
use_tools: bool):
|
| 192 |
"""Respond to query using specified configuration with optional MCP tools"""
|
| 193 |
|
| 194 |
+
# Always use the current UI values, regardless of preset selection
|
| 195 |
+
# Preset is only used for initializing UI fields, not for actual execution
|
| 196 |
+
llm_config_dict = create_custom_llm_config(
|
| 197 |
+
api_type, api_key, base_url, api_version, model, temperature, max_tokens
|
| 198 |
+
)
|
| 199 |
+
|
| 200 |
if preset_model != "Custom":
|
| 201 |
+
print(f"🔧 Using preset '{preset_model}' base with UI overrides:")
|
| 202 |
+
print(f" - API Type: {api_type}")
|
| 203 |
+
print(f" - Model: {model}")
|
| 204 |
+
print(f" - Base URL: {base_url}")
|
| 205 |
+
print(f" - API Version: {api_version}")
|
| 206 |
+
print(f" - Temperature: {temperature}")
|
| 207 |
+
print(f" - Max tokens: {max_tokens}")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 208 |
else:
|
|
|
|
|
|
|
|
|
|
|
|
|
| 209 |
print(f"🔧 Using custom configuration:")
|
| 210 |
+
print(f" - API Type: {api_type}")
|
| 211 |
+
print(f" - Model: {model}")
|
| 212 |
+
print(f" - Base URL: {base_url}")
|
| 213 |
+
print(f" - API Version: {api_version}")
|
| 214 |
print(f" - Temperature: {temperature}")
|
| 215 |
print(f" - Max tokens: {max_tokens}")
|
|
|
|
|
|
|
| 216 |
|
| 217 |
+
# Always rebuild brain with current UI configuration
|
| 218 |
print(f"🧠 Building brain with final config:")
|
| 219 |
+
print(f" - Final API type: {llm_config_dict['api_type']}")
|
| 220 |
+
print(f" - Final model: {llm_config_dict['model']}")
|
| 221 |
print(f" - Final temperature: {llm_config_dict['temperature']}")
|
| 222 |
print(f" - Final max_tokens: {llm_config_dict['max_tokens']}")
|
| 223 |
brain = build_brain_with_config(llm_config_dict)
|