Spaces:
Sleeping
Sleeping
modify prompt to return idrect answer
Browse files- Agents/agent.py +17 -3
- Prompts/{SystemPrompt copy.txt → SystemPrompt.txt} +0 -0
- app.py +2 -1
Agents/agent.py
CHANGED
|
@@ -46,10 +46,14 @@ if not GROQ_API_KEY:
|
|
| 46 |
|
| 47 |
langfuse_handler = CallbackHandler()
|
| 48 |
|
| 49 |
-
with open("Prompts/
|
| 50 |
systemPrompt = f.read()
|
| 51 |
print(f"systemPrompt {systemPrompt}")
|
| 52 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 53 |
sysMsg = SystemMessage(content=systemPrompt)
|
| 54 |
print(f"system message {sysMsg}")
|
| 55 |
|
|
@@ -213,7 +217,8 @@ def tools_condition1(
|
|
| 213 |
if hasattr(ai_message, "tool_calls") and len(ai_message.tool_calls) > 0:
|
| 214 |
print("tools")
|
| 215 |
return "tools"
|
| 216 |
-
return "
|
|
|
|
| 217 |
|
| 218 |
|
| 219 |
def build_graph(provider: str="groq"):
|
|
@@ -238,6 +243,12 @@ def build_graph(provider: str="groq"):
|
|
| 238 |
raise ValueError("Invalid Provider. Choose 'groq or 'huggingface'")
|
| 239 |
llmWithTools = llm.bind_tools(tools)
|
| 240 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 241 |
def assistant(state: MessagesState):
|
| 242 |
#msg
|
| 243 |
#msg = {"messages": [sysMsg] + state["messages"]}
|
|
@@ -264,13 +275,16 @@ def build_graph(provider: str="groq"):
|
|
| 264 |
#builder.add_node("retriever",retriever)
|
| 265 |
builder.add_node("assistant",assistant)
|
| 266 |
builder.add_node("tools", create_tool_node_with_fallback(tools))
|
|
|
|
| 267 |
builder.add_edge(START,"assistant")
|
| 268 |
|
| 269 |
builder.add_conditional_edges(
|
| 270 |
"assistant",
|
| 271 |
-
tools_condition
|
| 272 |
)
|
| 273 |
builder.add_edge("tools","assistant")
|
|
|
|
|
|
|
| 274 |
'''if(directPath):
|
| 275 |
builder.add_edge("assistant","tools")
|
| 276 |
builder.add_edge("tools",END)
|
|
|
|
| 46 |
|
| 47 |
langfuse_handler = CallbackHandler()
|
| 48 |
|
| 49 |
+
with open("Prompts/SystemPrompt.txt","r",encoding="utf-8") as f:
|
| 50 |
systemPrompt = f.read()
|
| 51 |
print(f"systemPrompt {systemPrompt}")
|
| 52 |
|
| 53 |
+
with open("Prompts/SystemPromptFinal.txt","r",encoding="utf-8") as f:
|
| 54 |
+
systemPromptFinalAnswer = f.read()
|
| 55 |
+
print(f"systemPromptFinalAnswer {systemPromptFinalAnswer}")
|
| 56 |
+
|
| 57 |
sysMsg = SystemMessage(content=systemPrompt)
|
| 58 |
print(f"system message {sysMsg}")
|
| 59 |
|
|
|
|
| 217 |
if hasattr(ai_message, "tool_calls") and len(ai_message.tool_calls) > 0:
|
| 218 |
print("tools")
|
| 219 |
return "tools"
|
| 220 |
+
return "ResponseParser"
|
| 221 |
+
#return "__end__"
|
| 222 |
|
| 223 |
|
| 224 |
def build_graph(provider: str="groq"):
|
|
|
|
| 243 |
raise ValueError("Invalid Provider. Choose 'groq or 'huggingface'")
|
| 244 |
llmWithTools = llm.bind_tools(tools)
|
| 245 |
|
| 246 |
+
def parseResponse(state:MessagesState):
|
| 247 |
+
finalContent = str(state["messages"][-1]["content"])
|
| 248 |
+
formattedMessage = [{"role": "system", "content": systemPromptFinalAnswer}] + \
|
| 249 |
+
[{"role":"user","content":finalContent} ]
|
| 250 |
+
return {"messages": [llmWithTools.invoke(formattedMessage)]}
|
| 251 |
+
|
| 252 |
def assistant(state: MessagesState):
|
| 253 |
#msg
|
| 254 |
#msg = {"messages": [sysMsg] + state["messages"]}
|
|
|
|
| 275 |
#builder.add_node("retriever",retriever)
|
| 276 |
builder.add_node("assistant",assistant)
|
| 277 |
builder.add_node("tools", create_tool_node_with_fallback(tools))
|
| 278 |
+
#builder.add_node("ResponseParser",parseResponse)
|
| 279 |
builder.add_edge(START,"assistant")
|
| 280 |
|
| 281 |
builder.add_conditional_edges(
|
| 282 |
"assistant",
|
| 283 |
+
tools_condition
|
| 284 |
)
|
| 285 |
builder.add_edge("tools","assistant")
|
| 286 |
+
#builder.add_edge("ResponseParser",END)
|
| 287 |
+
|
| 288 |
'''if(directPath):
|
| 289 |
builder.add_edge("assistant","tools")
|
| 290 |
builder.add_edge("tools",END)
|
Prompts/{SystemPrompt copy.txt → SystemPrompt.txt}
RENAMED
|
File without changes
|
app.py
CHANGED
|
@@ -75,12 +75,13 @@ class BasicAgent:
|
|
| 75 |
|
| 76 |
#self.graph.get_graph().draw_mermaid_png()
|
| 77 |
answer = messages['messages'][-1].content
|
|
|
|
| 78 |
#tools_calls = messages[-1].tool_calls
|
| 79 |
#print(f"tool calls : {str(tools_calls)}")
|
| 80 |
#print(messages[-1].__dict__)
|
| 81 |
#print(f"full answer : {answer}")
|
| 82 |
#print("Assistant Output:", messages[-1])
|
| 83 |
-
return answer
|
| 84 |
|
| 85 |
def run_and_submit_all( profile: gr.OAuthProfile | None):
|
| 86 |
"""
|
|
|
|
| 75 |
|
| 76 |
#self.graph.get_graph().draw_mermaid_png()
|
| 77 |
answer = messages['messages'][-1].content
|
| 78 |
+
print(f"ANWSER: {answer}")
|
| 79 |
#tools_calls = messages[-1].tool_calls
|
| 80 |
#print(f"tool calls : {str(tools_calls)}")
|
| 81 |
#print(messages[-1].__dict__)
|
| 82 |
#print(f"full answer : {answer}")
|
| 83 |
#print("Assistant Output:", messages[-1])
|
| 84 |
+
return answer
|
| 85 |
|
| 86 |
def run_and_submit_all( profile: gr.OAuthProfile | None):
|
| 87 |
"""
|