Commit ·
980918c
1
Parent(s): 5d748e2
Fixed streaming
Browse files- mainV2.py +31 -28
- models/system3.prompt +1 -1
- src/manager.py +47 -42
mainV2.py
CHANGED
|
@@ -11,48 +11,51 @@ if __name__ == "__main__":
|
|
| 11 |
|
| 12 |
model_manager = GeminiManager(toolsLoader=tool_loader, gemini_model="gemini-2.0-flash")
|
| 13 |
|
| 14 |
-
def respond(message, chat_history):
|
| 15 |
-
return model_manager.ask(message, chat_history)
|
| 16 |
-
|
| 17 |
def user_message(msg: str, history: list) -> tuple[str, list]:
|
| 18 |
"""Adds user message to chat history"""
|
| 19 |
history.append(gr.ChatMessage(role="user", content=msg))
|
| 20 |
return "", history
|
|
|
|
|
|
|
|
|
|
| 21 |
|
| 22 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 23 |
chatbot = gr.Chatbot(
|
| 24 |
avatar_images=("HASHIRU_2.png", "HASHIRU.png"),
|
| 25 |
-
type="messages"
|
|
|
|
|
|
|
|
|
|
| 26 |
)
|
| 27 |
-
input_box = gr.Textbox()
|
| 28 |
-
|
| 29 |
-
|
| 30 |
-
|
| 31 |
-
|
| 32 |
-
chat_history.append({
|
| 33 |
-
"role":"user",
|
| 34 |
-
"content":message
|
| 35 |
-
})
|
| 36 |
-
print("Chat history:", chat_history)
|
| 37 |
-
chat_history = model_manager.run(chat_history)
|
| 38 |
-
return "", chat_history
|
| 39 |
-
|
| 40 |
-
msg_store = gr.State("")
|
| 41 |
|
| 42 |
input_box.submit(
|
| 43 |
-
lambda msg: (msg, msg, ""), # Store message and clear input
|
| 44 |
-
inputs=[input_box],
|
| 45 |
-
outputs=[msg_store, input_box, input_box],
|
| 46 |
-
queue=False
|
| 47 |
-
).then(
|
| 48 |
user_message, # Add user message to chat
|
| 49 |
-
inputs=[
|
| 50 |
outputs=[input_box, chatbot],
|
| 51 |
-
queue=False
|
| 52 |
).then(
|
| 53 |
model_manager.run, # Generate and stream response
|
| 54 |
-
inputs=
|
| 55 |
-
outputs=chatbot
|
|
|
|
|
|
|
| 56 |
)
|
|
|
|
| 57 |
|
| 58 |
demo.launch(share=True)
|
|
|
|
| 11 |
|
| 12 |
model_manager = GeminiManager(toolsLoader=tool_loader, gemini_model="gemini-2.0-flash")
|
| 13 |
|
|
|
|
|
|
|
|
|
|
| 14 |
def user_message(msg: str, history: list) -> tuple[str, list]:
|
| 15 |
"""Adds user message to chat history"""
|
| 16 |
history.append(gr.ChatMessage(role="user", content=msg))
|
| 17 |
return "", history
|
| 18 |
+
|
| 19 |
+
def handle_undo(history, undo_data: gr.UndoData):
|
| 20 |
+
return history[:undo_data.index], history[undo_data.index]['content']
|
| 21 |
|
| 22 |
+
def handle_retry(history, retry_data: gr.RetryData):
|
| 23 |
+
new_history = history[:retry_data.index]
|
| 24 |
+
yield from model_manager.run(new_history)
|
| 25 |
+
|
| 26 |
+
def handle_edit(history, edit_data: gr.EditData):
|
| 27 |
+
new_history = history[:edit_data.index]
|
| 28 |
+
new_history[-1]['content'] = edit_data.value
|
| 29 |
+
return new_history
|
| 30 |
+
|
| 31 |
+
with gr.Blocks(fill_width=True, fill_height=True) as demo:
|
| 32 |
+
gr.Markdown("# Hashiru AI")
|
| 33 |
+
|
| 34 |
chatbot = gr.Chatbot(
|
| 35 |
avatar_images=("HASHIRU_2.png", "HASHIRU.png"),
|
| 36 |
+
type="messages",
|
| 37 |
+
show_copy_button=True,
|
| 38 |
+
editable="user",
|
| 39 |
+
scale=1
|
| 40 |
)
|
| 41 |
+
input_box = gr.Textbox(submit_btn=True, stop_btn=True, max_lines=5, label="Chat Message", scale=0)
|
| 42 |
+
|
| 43 |
+
chatbot.undo(handle_undo, chatbot, [chatbot, input_box])
|
| 44 |
+
chatbot.retry(handle_retry, chatbot, chatbot)
|
| 45 |
+
chatbot.edit(handle_edit, chatbot, chatbot)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 46 |
|
| 47 |
input_box.submit(
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 48 |
user_message, # Add user message to chat
|
| 49 |
+
inputs=[input_box, chatbot],
|
| 50 |
outputs=[input_box, chatbot],
|
| 51 |
+
queue=False,
|
| 52 |
).then(
|
| 53 |
model_manager.run, # Generate and stream response
|
| 54 |
+
inputs=chatbot,
|
| 55 |
+
outputs=chatbot,
|
| 56 |
+
show_progress="full",
|
| 57 |
+
trigger_mode="always_last"
|
| 58 |
)
|
| 59 |
+
input_box.submit(lambda: "", None, [input_box])
|
| 60 |
|
| 61 |
demo.launch(share=True)
|
models/system3.prompt
CHANGED
|
@@ -103,7 +103,7 @@ Strictly follow the schema required for invoking the tools and agents. Do not de
|
|
| 103 |
</Rule>
|
| 104 |
|
| 105 |
<Rule>
|
| 106 |
-
Once you have the answer, provide it to the user in a clear and concise manner
|
| 107 |
</Rule>
|
| 108 |
|
| 109 |
<Rule>
|
|
|
|
| 103 |
</Rule>
|
| 104 |
|
| 105 |
<Rule>
|
| 106 |
+
Once you have the answer, provide it to the user in a clear and concise manner.
|
| 107 |
</Rule>
|
| 108 |
|
| 109 |
<Rule>
|
src/manager.py
CHANGED
|
@@ -16,6 +16,7 @@ handler = logging.StreamHandler(sys.stdout)
|
|
| 16 |
# handler.setLevel(logging.DEBUG)
|
| 17 |
logger.addHandler(handler)
|
| 18 |
|
|
|
|
| 19 |
class GeminiManager:
|
| 20 |
def __init__(self, toolsLoader: ToolLoader, system_prompt_file="./models/system3.prompt", gemini_model="gemini-2.5-pro-exp-03-25"):
|
| 21 |
load_dotenv()
|
|
@@ -30,22 +31,24 @@ class GeminiManager:
|
|
| 30 |
|
| 31 |
def generate_response(self, messages):
|
| 32 |
return self.client.models.generate_content(
|
| 33 |
-
|
| 34 |
-
|
| 35 |
-
|
| 36 |
-
|
| 37 |
-
|
| 38 |
-
|
| 39 |
-
|
| 40 |
-
|
| 41 |
-
|
| 42 |
def handle_tool_calls(self, response):
|
| 43 |
parts = []
|
| 44 |
for function_call in response.function_calls:
|
| 45 |
toolResponse = None
|
| 46 |
-
logger.info(
|
|
|
|
| 47 |
try:
|
| 48 |
-
toolResponse = self.toolsLoader.runTool(
|
|
|
|
| 49 |
except Exception as e:
|
| 50 |
logger.warning(f"Error running tool: {e}")
|
| 51 |
toolResponse = {
|
|
@@ -55,53 +58,57 @@ class GeminiManager:
|
|
| 55 |
}
|
| 56 |
logger.debug(f"Tool Response: {toolResponse}")
|
| 57 |
tool_content = types.Part.from_function_response(
|
| 58 |
-
|
| 59 |
-
|
| 60 |
try:
|
| 61 |
self.toolsLoader.load_tools()
|
| 62 |
except Exception as e:
|
| 63 |
logger.info(f"Error loading tools: {e}. Deleting the tool.")
|
| 64 |
# delete the created tool
|
| 65 |
-
self.toolsLoader.delete_tool(
|
|
|
|
| 66 |
tool_content = types.Part.from_function_response(
|
| 67 |
-
|
| 68 |
-
|
| 69 |
parts.append(tool_content)
|
| 70 |
return {
|
| 71 |
-
|
| 72 |
-
|
| 73 |
role='model' if self.model_name == "gemini-2.5-pro-exp-03-25" else 'tool',
|
| 74 |
parts=parts
|
| 75 |
-
|
| 76 |
-
|
| 77 |
-
|
| 78 |
def format_chat_history(self, messages=[]):
|
| 79 |
formatted_history = []
|
| 80 |
for message in messages:
|
| 81 |
# Skip thinking messages (messages with metadata)
|
| 82 |
if not (message.get("role") == "assistant" and "metadata" in message):
|
| 83 |
role = "model"
|
| 84 |
-
parts=[types.Part.from_text(text=message.get("content", ""))]
|
| 85 |
match message.get("role"):
|
| 86 |
case "user":
|
| 87 |
role = "user"
|
| 88 |
case "tool":
|
| 89 |
role = "tool"
|
| 90 |
-
formatted_history.append(
|
|
|
|
| 91 |
continue
|
| 92 |
case "function_call":
|
| 93 |
role = "model"
|
| 94 |
-
formatted_history.append(
|
|
|
|
| 95 |
continue
|
| 96 |
case _:
|
| 97 |
-
role = "model"
|
| 98 |
formatted_history.append(types.Content(
|
| 99 |
role=role,
|
| 100 |
parts=parts
|
| 101 |
))
|
| 102 |
return formatted_history
|
| 103 |
-
|
| 104 |
-
def run(self,
|
|
|
|
| 105 |
chat_history = self.format_chat_history(messages)
|
| 106 |
logger.debug(f"Chat history: {chat_history}")
|
| 107 |
try:
|
|
@@ -109,18 +116,19 @@ class GeminiManager:
|
|
| 109 |
except Exception as e:
|
| 110 |
logger.debug(f"Error generating response: {e}")
|
| 111 |
messages.append({
|
| 112 |
-
"role":"assistant",
|
| 113 |
-
"content":f"Error generating response: {e}"
|
| 114 |
})
|
| 115 |
logger.error(f"Error generating response: {e}")
|
| 116 |
return messages
|
| 117 |
logger.debug(f"Response: {response}")
|
| 118 |
-
|
|
|
|
| 119 |
if (not response.text and not response.function_calls):
|
| 120 |
messages.append({
|
| 121 |
-
"role":"assistant",
|
| 122 |
-
"content":"No response from the model.",
|
| 123 |
-
"metadata":{"title":"No response from the model."}
|
| 124 |
})
|
| 125 |
|
| 126 |
# Attach the llm response to the messages
|
|
@@ -130,22 +138,19 @@ class GeminiManager:
|
|
| 130 |
"content": response.text
|
| 131 |
})
|
| 132 |
yield messages
|
| 133 |
-
|
| 134 |
# Attach the function call response to the messages
|
| 135 |
if response.candidates[0].content and response.candidates[0].content.parts:
|
| 136 |
# messages.append(response.candidates[0].content)
|
| 137 |
messages.append({
|
| 138 |
-
"role":"function_call",
|
| 139 |
"content": repr(response.candidates[0].content),
|
| 140 |
})
|
| 141 |
-
|
| 142 |
-
|
| 143 |
# Invoke the function calls if any and attach the response to the messages
|
| 144 |
if response.function_calls:
|
| 145 |
calls = self.handle_tool_calls(response)
|
| 146 |
messages.append(calls)
|
| 147 |
-
yield messages
|
| 148 |
-
|
| 149 |
-
|
| 150 |
-
return
|
| 151 |
-
return messages
|
|
|
|
| 16 |
# handler.setLevel(logging.DEBUG)
|
| 17 |
logger.addHandler(handler)
|
| 18 |
|
| 19 |
+
|
| 20 |
class GeminiManager:
|
| 21 |
def __init__(self, toolsLoader: ToolLoader, system_prompt_file="./models/system3.prompt", gemini_model="gemini-2.5-pro-exp-03-25"):
|
| 22 |
load_dotenv()
|
|
|
|
| 31 |
|
| 32 |
def generate_response(self, messages):
|
| 33 |
return self.client.models.generate_content(
|
| 34 |
+
model=self.model_name,
|
| 35 |
+
contents=messages,
|
| 36 |
+
config=types.GenerateContentConfig(
|
| 37 |
+
system_instruction=self.system_prompt,
|
| 38 |
+
temperature=0.2,
|
| 39 |
+
tools=self.toolsLoader.getTools(),
|
| 40 |
+
),
|
| 41 |
+
)
|
| 42 |
+
|
| 43 |
def handle_tool_calls(self, response):
|
| 44 |
parts = []
|
| 45 |
for function_call in response.function_calls:
|
| 46 |
toolResponse = None
|
| 47 |
+
logger.info(
|
| 48 |
+
f"Function Name: {function_call.name}, Arguments: {function_call.args}")
|
| 49 |
try:
|
| 50 |
+
toolResponse = self.toolsLoader.runTool(
|
| 51 |
+
function_call.name, function_call.args)
|
| 52 |
except Exception as e:
|
| 53 |
logger.warning(f"Error running tool: {e}")
|
| 54 |
toolResponse = {
|
|
|
|
| 58 |
}
|
| 59 |
logger.debug(f"Tool Response: {toolResponse}")
|
| 60 |
tool_content = types.Part.from_function_response(
|
| 61 |
+
name=function_call.name,
|
| 62 |
+
response={"result": toolResponse})
|
| 63 |
try:
|
| 64 |
self.toolsLoader.load_tools()
|
| 65 |
except Exception as e:
|
| 66 |
logger.info(f"Error loading tools: {e}. Deleting the tool.")
|
| 67 |
# delete the created tool
|
| 68 |
+
self.toolsLoader.delete_tool(
|
| 69 |
+
toolResponse['output']['tool_name'], toolResponse['output']['tool_file_path'])
|
| 70 |
tool_content = types.Part.from_function_response(
|
| 71 |
+
name=function_call.name,
|
| 72 |
+
response={"result": f"{function_call.name} with {function_call.args} doesn't follow the required format, please read the other tool implementations for reference." + str(e)})
|
| 73 |
parts.append(tool_content)
|
| 74 |
return {
|
| 75 |
+
"role": "tool",
|
| 76 |
+
"content": repr(types.Content(
|
| 77 |
role='model' if self.model_name == "gemini-2.5-pro-exp-03-25" else 'tool',
|
| 78 |
parts=parts
|
| 79 |
+
))
|
| 80 |
+
}
|
| 81 |
+
|
| 82 |
def format_chat_history(self, messages=[]):
|
| 83 |
formatted_history = []
|
| 84 |
for message in messages:
|
| 85 |
# Skip thinking messages (messages with metadata)
|
| 86 |
if not (message.get("role") == "assistant" and "metadata" in message):
|
| 87 |
role = "model"
|
| 88 |
+
parts = [types.Part.from_text(text=message.get("content", ""))]
|
| 89 |
match message.get("role"):
|
| 90 |
case "user":
|
| 91 |
role = "user"
|
| 92 |
case "tool":
|
| 93 |
role = "tool"
|
| 94 |
+
formatted_history.append(
|
| 95 |
+
eval(message.get("content", "")))
|
| 96 |
continue
|
| 97 |
case "function_call":
|
| 98 |
role = "model"
|
| 99 |
+
formatted_history.append(
|
| 100 |
+
eval(message.get("content", "")))
|
| 101 |
continue
|
| 102 |
case _:
|
| 103 |
+
role = "model"
|
| 104 |
formatted_history.append(types.Content(
|
| 105 |
role=role,
|
| 106 |
parts=parts
|
| 107 |
))
|
| 108 |
return formatted_history
|
| 109 |
+
|
| 110 |
+
def run(self, messages):
|
| 111 |
+
print("Messages: ", messages)
|
| 112 |
chat_history = self.format_chat_history(messages)
|
| 113 |
logger.debug(f"Chat history: {chat_history}")
|
| 114 |
try:
|
|
|
|
| 116 |
except Exception as e:
|
| 117 |
logger.debug(f"Error generating response: {e}")
|
| 118 |
messages.append({
|
| 119 |
+
"role": "assistant",
|
| 120 |
+
"content": f"Error generating response: {e}"
|
| 121 |
})
|
| 122 |
logger.error(f"Error generating response: {e}")
|
| 123 |
return messages
|
| 124 |
logger.debug(f"Response: {response}")
|
| 125 |
+
print("Response: ", response)
|
| 126 |
+
|
| 127 |
if (not response.text and not response.function_calls):
|
| 128 |
messages.append({
|
| 129 |
+
"role": "assistant",
|
| 130 |
+
"content": "No response from the model.",
|
| 131 |
+
"metadata": {"title": "No response from the model."}
|
| 132 |
})
|
| 133 |
|
| 134 |
# Attach the llm response to the messages
|
|
|
|
| 138 |
"content": response.text
|
| 139 |
})
|
| 140 |
yield messages
|
| 141 |
+
|
| 142 |
# Attach the function call response to the messages
|
| 143 |
if response.candidates[0].content and response.candidates[0].content.parts:
|
| 144 |
# messages.append(response.candidates[0].content)
|
| 145 |
messages.append({
|
| 146 |
+
"role": "function_call",
|
| 147 |
"content": repr(response.candidates[0].content),
|
| 148 |
})
|
| 149 |
+
|
|
|
|
| 150 |
# Invoke the function calls if any and attach the response to the messages
|
| 151 |
if response.function_calls:
|
| 152 |
calls = self.handle_tool_calls(response)
|
| 153 |
messages.append(calls)
|
| 154 |
+
yield from self.run(messages)
|
| 155 |
+
print("Final messages: ", messages)
|
| 156 |
+
return messages
|
|
|
|
|
|