trygithubactions / my_agent /utils /business_interaction.py
subashpoudel's picture
Implemented the business interaction after introduction
3f79d7a
raw
history blame
1.91 kB
import os
from langchain_groq import ChatGroq
from langgraph.graph import StateGraph, MessagesState, START, END
from langgraph.checkpoint.memory import MemorySaver
from langchain_core.messages import SystemMessage
from pydantic import BaseModel, ConfigDict, Field
from typing import Optional, List
from .models_loader import llm
from .prompts import business_interaction_prompt
# State model
class State(BaseModel):
interactions: Optional[list] = []
model_config = ConfigDict(arbitrary_types_allowed=True)
# Global business state (shared)
business_state = State()
class BusinessInteractionChatbot2:
def __init__(self):
self.memory = MemorySaver()
# self.llm = ChatGroq(model_name="Gemma2-9b-It")
self.llm = llm
self.workflow = self._initialize_workflow()
self.interact_agent = self.workflow.compile(checkpointer=self.memory)
self.messages = []
def _initialize_workflow(self):
workflow = StateGraph(MessagesState)
workflow.add_node("chatbot", self._call_model)
workflow.add_edge(START, "chatbot")
workflow.add_edge("chatbot", END)
return workflow
def _call_model(self, state):
template = business_interaction_prompt
messages = [SystemMessage(content=template)] + state["messages"]
response = self.llm.invoke(messages)
return {"messages": [response]}
def chat(self, user_input: str , business_details: dict={}):
self.messages.append({"role": "user", "content": user_input})
config = {"configurable": {"thread_id": "1"}}
response = self.interact_agent.invoke({"messages": [user_input + str(business_details)]}, config)['messages'][-1].content
self.messages.append({"role": "assistant", "content": response})
business_state.interactions.append({'user': user_input, 'agent_response': response})
return response