Spaces:
Running
Running
| from langchain_core.runnables import Runnable | |
| from cuga.backend.cuga_graph.nodes.shared.base_agent import BaseAgent | |
| from cuga.backend.llm.models import LLMManager | |
| from cuga.backend.llm.utils.helpers import load_prompt_simple | |
| llm_manager = LLMManager() | |
| def reflection_task(llm, enable_format=False) -> Runnable: | |
| prompt_template = load_prompt_simple( | |
| "./prompts/reflection_system.jinja2", | |
| "./prompts/reflection_user.jinja2", | |
| ) | |
| return BaseAgent.get_chain(prompt_template, llm, wx_json_mode="no_format") | |