fmarky's picture
style: replace specific char in app
bec8a1d verified
#!/usr/bin/env python3
import os
import asyncio
import gradio as gr
from dotenv import load_dotenv
from typing import Annotated, TypedDict
from langgraph.graph import StateGraph, START
from langgraph.prebuilt import ToolNode, tools_condition
from langgraph.graph.message import add_messages
from langchain_core.messages import HumanMessage, SystemMessage, AnyMessage
from langchain_mcp_adapters.resources import load_mcp_resources
from langchain_mcp_adapters.tools import load_mcp_tools
from langchain_mcp_adapters.prompts import load_mcp_prompt
from langchain_google_genai import ChatGoogleGenerativeAI
from mcp import ClientSession
from mcp.client.streamable_http import streamablehttp_client
load_dotenv()
MCP_URL = os.getenv("MCP_SERVER_URL", "http://localhost:7860/mcp")
PROMPT_NAME = os.getenv("MCP_PROMPT_NAME", "ship_meme_for_commit")
class AgentState(TypedDict):
messages: Annotated[list[AnyMessage], add_messages]
async def run_for_commit(commit_id: str) -> str:
async with streamablehttp_client(MCP_URL) as (read, write, _):
async with ClientSession(read, write) as session:
await session.initialize()
# 1) Resource: gitdiff://{commit_id}
resources = await load_mcp_resources(
session, uris=[f"gitdiff://{commit_id}"]
)
git_diff_content = "".join(
(r.data for r in resources if isinstance(getattr(r, "data", None), str))
)
# 2) Tools (bound to this live session)
tools = await load_mcp_tools(session)
# 3) Prompt
prompts = await load_mcp_prompt(
session,
PROMPT_NAME,
arguments={
"commit_id": commit_id,
"git_diff_content": git_diff_content,
},
)
sys_text = (
prompts[0].content
if prompts
else "You are an agent. Use tools as needed."
)
# 4) LLM + tools
llm = ChatGoogleGenerativeAI(model="gemini-2.5-flash", temperature=0)
llm_with_tools = llm.bind_tools(tools)
# 5) Assistant node (async)
async def assistant(state: AgentState):
sys_msg = SystemMessage(content=sys_text)
msg = await llm_with_tools.ainvoke([sys_msg] + state["messages"])
return {"messages": [msg]}
# 6) Graph compile + run inside the same context
g = StateGraph(AgentState)
g.add_node("assistant", assistant)
g.add_node("tools", ToolNode(tools))
g.add_edge(START, "assistant")
g.add_conditional_edges("assistant", tools_condition)
agent = g.compile()
result = await agent.ainvoke(
{
"messages": [
HumanMessage(content=f"Ship a meme for commit {commit_id}")
]
}
)
return result["messages"][-1].content
# -------- Gradio --------
def ui_fn(message, _history):
commit = (message or "").strip() or "demo-42"
try:
return asyncio.run(run_for_commit(commit))
except Exception as e:
return f"Client error: {type(e).__name__}: {e}"
demo = gr.ChatInterface(
fn=ui_fn,
type="messages",
title="MemeOps - MCP + LangGraph",
description="Give your commit ID.",
examples=["demo-42"],
)
if __name__ == "__main__":
demo.launch(server_name="0.0.0.0", server_port=int(os.getenv("PORT", "7860")))