Spaces:
Sleeping
Sleeping
Update agent.py
Browse files
agent.py
CHANGED
|
@@ -47,12 +47,18 @@ sys_msg = SystemMessage(content=system_prompt)
|
|
| 47 |
|
| 48 |
# Loading the assistant
|
| 49 |
#chat = ChatGoogleGenerativeAI(model="gemini-2.0-flash", temperature=0)
|
| 50 |
-
chat = ChatGoogleGenerativeAI(
|
| 51 |
-
|
| 52 |
-
|
| 53 |
-
|
| 54 |
-
|
| 55 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 56 |
),
|
| 57 |
)
|
| 58 |
|
|
@@ -82,30 +88,13 @@ def simple_graph():
|
|
| 82 |
def assistant(state: MessagesState):
|
| 83 |
"""Assistant node"""
|
| 84 |
return {"messages": [chat_with_tools.invoke([sys_msg] + state["messages"])]}
|
| 85 |
-
|
| 86 |
-
# def retriever(state: MessagesState):
|
| 87 |
-
# """Retriever node"""
|
| 88 |
-
# similar_question = vector_store.similarity_search(state["messages"][0].content)
|
| 89 |
-
|
| 90 |
-
# if similar_question: # Check if the list is not empty
|
| 91 |
-
# example_msg = HumanMessage(
|
| 92 |
-
# content=f"Here I provide a similar question and answer for reference: \n\n{similar_question[0].page_content}",
|
| 93 |
-
# )
|
| 94 |
-
# return {"messages": [sys_msg] + state["messages"] + [example_msg]}
|
| 95 |
-
# else:
|
| 96 |
-
# # Handle the case when no similar questions are found
|
| 97 |
-
# return {"messages": [sys_msg] + state["messages"]}
|
| 98 |
-
|
| 99 |
|
| 100 |
# Build graph / nodes
|
| 101 |
builder = StateGraph(MessagesState)
|
| 102 |
-
#builder.add_node("retriever", retriever) # Retriever
|
| 103 |
builder.add_node("assistant", assistant) # Assistant
|
| 104 |
builder.add_node("tools", ToolNode(tools)) # Tools
|
| 105 |
|
| 106 |
# Logic / edges
|
| 107 |
-
# builder.add_edge(START, "retriever")
|
| 108 |
-
# builder.add_edge("retriever", "assistant")
|
| 109 |
builder.add_edge(START, "assistant")
|
| 110 |
builder.add_conditional_edges("assistant", tools_condition)
|
| 111 |
builder.add_edge("tools", "assistant")
|
|
|
|
| 47 |
|
| 48 |
# Loading the assistant
|
| 49 |
#chat = ChatGoogleGenerativeAI(model="gemini-2.0-flash", temperature=0)
|
| 50 |
+
# chat = ChatGoogleGenerativeAI(
|
| 51 |
+
# model="gemini-2.5-flash-preview-04-17",
|
| 52 |
+
# temperature=0,
|
| 53 |
+
# max_retries=1,
|
| 54 |
+
# rate_limiter=InMemoryRateLimiter(
|
| 55 |
+
# requests_per_second=1/60,
|
| 56 |
+
# ),
|
| 57 |
+
# )
|
| 58 |
+
chat = ChatGroq(model_name="llama3-70b-8192",
|
| 59 |
+
temperature=0,
|
| 60 |
+
rate_limiter=InMemoryRateLimiter(
|
| 61 |
+
requests_per_second=1/60,
|
| 62 |
),
|
| 63 |
)
|
| 64 |
|
|
|
|
| 88 |
def assistant(state: MessagesState):
|
| 89 |
"""Assistant node"""
|
| 90 |
return {"messages": [chat_with_tools.invoke([sys_msg] + state["messages"])]}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 91 |
|
| 92 |
# Build graph / nodes
|
| 93 |
builder = StateGraph(MessagesState)
|
|
|
|
| 94 |
builder.add_node("assistant", assistant) # Assistant
|
| 95 |
builder.add_node("tools", ToolNode(tools)) # Tools
|
| 96 |
|
| 97 |
# Logic / edges
|
|
|
|
|
|
|
| 98 |
builder.add_edge(START, "assistant")
|
| 99 |
builder.add_conditional_edges("assistant", tools_condition)
|
| 100 |
builder.add_edge("tools", "assistant")
|