Spaces:
Sleeping
Sleeping
Update agent.py
Browse filesupdate retriever remove supabase because of some errors
agent.py
CHANGED
|
@@ -174,50 +174,23 @@ def build_graph(provider: str = "groq"):
|
|
| 174 |
llm_with_tools = llm.bind_tools(tools)
|
| 175 |
|
| 176 |
# Node
|
|
|
|
| 177 |
def assistant(state: MessagesState):
|
| 178 |
"""Assistant node"""
|
| 179 |
return {"messages": [llm_with_tools.invoke(state["messages"])]}
|
| 180 |
|
| 181 |
-
#
|
| 182 |
-
# """Retriever node"""
|
| 183 |
-
# similar_question = vector_store.similarity_search(state["messages"][0].content)
|
| 184 |
-
#example_msg = HumanMessage(
|
| 185 |
-
# content=f"Here I provide a similar question and answer for reference: \n\n{similar_question[0].page_content}",
|
| 186 |
-
# )
|
| 187 |
-
# return {"messages": [sys_msg] + state["messages"] + [example_msg]}
|
| 188 |
-
|
| 189 |
-
from langchain_core.messages import AIMessage
|
| 190 |
-
|
| 191 |
-
def retriever(state: MessagesState):
|
| 192 |
-
query = state["messages"][-1].content
|
| 193 |
-
similar_doc = vector_store.similarity_search(query, k=1)[0]
|
| 194 |
-
|
| 195 |
-
content = similar_doc.page_content
|
| 196 |
-
if "Final answer :" in content:
|
| 197 |
-
answer = content.split("Final answer :")[-1].strip()
|
| 198 |
-
else:
|
| 199 |
-
answer = content.strip()
|
| 200 |
-
|
| 201 |
-
return {"messages": [AIMessage(content=answer)]}
|
| 202 |
-
|
| 203 |
-
# builder = StateGraph(MessagesState)
|
| 204 |
-
#builder.add_node("retriever", retriever)
|
| 205 |
-
#builder.add_node("assistant", assistant)
|
| 206 |
-
#builder.add_node("tools", ToolNode(tools))
|
| 207 |
-
#builder.add_edge(START, "retriever")
|
| 208 |
-
#builder.add_edge("retriever", "assistant")
|
| 209 |
-
#builder.add_conditional_edges(
|
| 210 |
-
# "assistant",
|
| 211 |
-
# tools_condition,
|
| 212 |
-
#)
|
| 213 |
-
#builder.add_edge("tools", "assistant")
|
| 214 |
-
|
| 215 |
builder = StateGraph(MessagesState)
|
| 216 |
-
builder.add_node("
|
| 217 |
-
|
| 218 |
-
|
| 219 |
-
builder.set_entry_point(
|
| 220 |
-
builder.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 221 |
|
| 222 |
# Compile graph
|
| 223 |
-
return builder.compile()
|
|
|
|
| 174 |
llm_with_tools = llm.bind_tools(tools)
|
| 175 |
|
| 176 |
# Node
|
| 177 |
+
# Node
|
| 178 |
def assistant(state: MessagesState):
|
| 179 |
"""Assistant node"""
|
| 180 |
return {"messages": [llm_with_tools.invoke(state["messages"])]}
|
| 181 |
|
| 182 |
+
# Skip the retriever and just use the LLM directly
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 183 |
builder = StateGraph(MessagesState)
|
| 184 |
+
builder.add_node("assistant", assistant)
|
| 185 |
+
builder.add_node("tools", ToolNode(tools))
|
| 186 |
+
|
| 187 |
+
builder.set_entry_point(START)
|
| 188 |
+
builder.add_edge(START, "assistant")
|
| 189 |
+
builder.add_conditional_edges(
|
| 190 |
+
"assistant",
|
| 191 |
+
tools_condition,
|
| 192 |
+
)
|
| 193 |
+
builder.add_edge("tools", "assistant")
|
| 194 |
|
| 195 |
# Compile graph
|
| 196 |
+
return builder.compile()
|