Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -64,11 +64,10 @@ class AdvancedAgent:
|
|
| 64 |
sys_msg = SystemMessage(content=system_prompt)
|
| 65 |
llm_with_tools = llm.bind_tools(tools)
|
| 66 |
|
| 67 |
-
# Assistant node must be able to access llm_with_tools.
|
| 68 |
def assistant(state: MessagesState):
|
| 69 |
"""Assistant node that processes messages and generates responses"""
|
| 70 |
messages = state["messages"]
|
| 71 |
-
response = llm_with_tools.invoke(messages)
|
| 72 |
return {"messages": messages + [response]}
|
| 73 |
|
| 74 |
tools_node = ToolNode(tools)
|
|
@@ -77,13 +76,10 @@ class AdvancedAgent:
|
|
| 77 |
builder.add_node("assistant", assistant)
|
| 78 |
builder.add_node("tools", tools_node)
|
| 79 |
|
| 80 |
-
# The correct flow is: assistant -> tools (if tool needed) -> assistant (after tool used) -> END (if done)
|
| 81 |
builder.set_entry_point("assistant")
|
| 82 |
builder.add_edge("assistant", "tools")
|
| 83 |
builder.add_edge("tools", "assistant")
|
| 84 |
builder.add_edge("assistant", END)
|
| 85 |
-
|
| 86 |
-
# Conditional: if tools_condition is met, go to tools, else END
|
| 87 |
builder.add_conditional_edges(
|
| 88 |
"assistant",
|
| 89 |
tools_condition,
|
|
@@ -99,7 +95,7 @@ class AdvancedAgent:
|
|
| 99 |
HumanMessage(content=question)
|
| 100 |
]
|
| 101 |
try:
|
| 102 |
-
result = self.graph.invoke(
|
| 103 |
final_messages = result["messages"]
|
| 104 |
ai_messages = [msg for msg in final_messages if isinstance(msg, AIMessage)]
|
| 105 |
if not ai_messages:
|
|
@@ -262,4 +258,4 @@ if __name__ == "__main__":
|
|
| 262 |
print("ℹ️ SPACE_ID environment variable not found (running locally?). Repo URL cannot be determined.")
|
| 263 |
print("-"*(60 + len(" App Starting ")) + "\n")
|
| 264 |
print("Launching Gradio Interface for Advanced Agent Evaluation...")
|
| 265 |
-
demo.launch(debug=True, share=False)
|
|
|
|
| 64 |
sys_msg = SystemMessage(content=system_prompt)
|
| 65 |
llm_with_tools = llm.bind_tools(tools)
|
| 66 |
|
|
|
|
| 67 |
def assistant(state: MessagesState):
|
| 68 |
"""Assistant node that processes messages and generates responses"""
|
| 69 |
messages = state["messages"]
|
| 70 |
+
response = llm_with_tools.invoke(messages) # <-- messages, not dict!
|
| 71 |
return {"messages": messages + [response]}
|
| 72 |
|
| 73 |
tools_node = ToolNode(tools)
|
|
|
|
| 76 |
builder.add_node("assistant", assistant)
|
| 77 |
builder.add_node("tools", tools_node)
|
| 78 |
|
|
|
|
| 79 |
builder.set_entry_point("assistant")
|
| 80 |
builder.add_edge("assistant", "tools")
|
| 81 |
builder.add_edge("tools", "assistant")
|
| 82 |
builder.add_edge("assistant", END)
|
|
|
|
|
|
|
| 83 |
builder.add_conditional_edges(
|
| 84 |
"assistant",
|
| 85 |
tools_condition,
|
|
|
|
| 95 |
HumanMessage(content=question)
|
| 96 |
]
|
| 97 |
try:
|
| 98 |
+
result = self.graph.invoke(messages) # <-- messages, not dict!
|
| 99 |
final_messages = result["messages"]
|
| 100 |
ai_messages = [msg for msg in final_messages if isinstance(msg, AIMessage)]
|
| 101 |
if not ai_messages:
|
|
|
|
| 258 |
print("ℹ️ SPACE_ID environment variable not found (running locally?). Repo URL cannot be determined.")
|
| 259 |
print("-"*(60 + len(" App Starting ")) + "\n")
|
| 260 |
print("Launching Gradio Interface for Advanced Agent Evaluation...")
|
| 261 |
+
demo.launch(debug=True, share=False)
|