Spaces:
Configuration error
Configuration error
oremaz
commited on
Commit
·
d294d73
1
Parent(s):
18ff5c5
Update agent.py
Browse files
agent.py
CHANGED
|
@@ -463,38 +463,23 @@ class EnhancedGAIAAgent:
|
|
| 463 |
tools=[analysis_tool, research_tool, code_tool]
|
| 464 |
)
|
| 465 |
|
| 466 |
-
def solve_gaia_question(self, question_data: Dict[str, Any]) -> str:
|
| 467 |
question = question_data.get("Question", "")
|
| 468 |
task_id = question_data.get("task_id", "")
|
| 469 |
-
|
| 470 |
context_prompt = f"""
|
| 471 |
-
|
| 472 |
-
|
| 473 |
-
|
| 474 |
-
|
| 475 |
-
|
| 476 |
-
|
| 477 |
-
|
| 478 |
-
|
| 479 |
-
|
| 480 |
-
Begin your reasoning process:
|
| 481 |
-
"""
|
| 482 |
-
|
| 483 |
try:
|
| 484 |
-
import asyncio
|
| 485 |
from llama_index.core.workflow import Context
|
| 486 |
-
|
| 487 |
-
# Créer le contexte
|
| 488 |
ctx = Context(self.coordinator)
|
| 489 |
-
|
| 490 |
-
# Fonction asynchrone pour exécuter l'agent
|
| 491 |
-
async def run_agent():
|
| 492 |
-
response = await self.coordinator.run(ctx=ctx, user_msg=context_prompt)
|
| 493 |
-
return response
|
| 494 |
-
|
| 495 |
-
# Exécuter de manière asynchrone
|
| 496 |
-
response = asyncio.run(run_agent())
|
| 497 |
return str(response)
|
| 498 |
-
|
| 499 |
except Exception as e:
|
| 500 |
return f"Error processing question: {str(e)}"
|
|
|
|
| 463 |
tools=[analysis_tool, research_tool, code_tool]
|
| 464 |
)
|
| 465 |
|
| 466 |
+
async def solve_gaia_question(self, question_data: Dict[str, Any]) -> str:
|
| 467 |
question = question_data.get("Question", "")
|
| 468 |
task_id = question_data.get("task_id", "")
|
|
|
|
| 469 |
context_prompt = f"""
|
| 470 |
+
GAIA Task ID: {task_id}
|
| 471 |
+
Question: {question}
|
| 472 |
+
{f"Associated files: {question_data.get('file_name', '')}" if 'file_name' in question_data else 'No files provided'}
|
| 473 |
+
Instructions:
|
| 474 |
+
1. Analyze this GAIA question using ReAct reasoning
|
| 475 |
+
2. Use specialist tools ONLY when their specific expertise is needed
|
| 476 |
+
3. Provide a precise, exact answer in GAIA format
|
| 477 |
+
Begin your reasoning process:
|
| 478 |
+
"""
|
|
|
|
|
|
|
|
|
|
| 479 |
try:
|
|
|
|
| 480 |
from llama_index.core.workflow import Context
|
|
|
|
|
|
|
| 481 |
ctx = Context(self.coordinator)
|
| 482 |
+
response = await self.coordinator.run(ctx=ctx, input=context_prompt)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 483 |
return str(response)
|
|
|
|
| 484 |
except Exception as e:
|
| 485 |
return f"Error processing question: {str(e)}"
|