File size: 557 Bytes
4851501
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
from typing import Dict, Any
from backend.services.executor import QueryExecutor

class OrchestratorAgent:
    def __init__(self):
        self.executor = QueryExecutor()

    async def process_query(self, query: str, history: list[Dict[str, str]] = None, model: str = None) -> Dict[str, Any]:
        """
        Delegates to QueryExecutor. Model param can be used to configure LLM if needed.
        """
        # For now, we rely on the default configured in LLMGateway
        return await self.executor.process_query_with_context(query, history or [])