Spaces:
Sleeping
Sleeping
Update LangGraphAgent to use z-ai/glm-4.5-air model instead of mistralai/mistral-small-3.2-24b-instruct
Browse files
agent.py
CHANGED
|
@@ -78,8 +78,7 @@ class LangGraphAgent:
|
|
| 78 |
# model="gemma-3-27b-it",
|
| 79 |
# )
|
| 80 |
model = ChatOpenRouter(
|
| 81 |
-
model="
|
| 82 |
-
|
| 83 |
)
|
| 84 |
if model is None and ChatOpenAI is not None:
|
| 85 |
model = ChatOpenAI(
|
|
@@ -129,7 +128,7 @@ class LangGraphAgent:
|
|
| 129 |
|
| 130 |
def __call__(self, question: str) -> str:
|
| 131 |
state = {"messages": [HumanMessage(content=question)]}
|
| 132 |
-
result = self.graph.invoke(state)
|
| 133 |
messages = result.get("messages", [])
|
| 134 |
# Return only the content after "FINAL ANSWER:"
|
| 135 |
for msg in reversed(messages):
|
|
|
|
| 78 |
# model="gemma-3-27b-it",
|
| 79 |
# )
|
| 80 |
model = ChatOpenRouter(
|
| 81 |
+
model="z-ai/glm-4.5-air:free",
|
|
|
|
| 82 |
)
|
| 83 |
if model is None and ChatOpenAI is not None:
|
| 84 |
model = ChatOpenAI(
|
|
|
|
| 128 |
|
| 129 |
def __call__(self, question: str) -> str:
|
| 130 |
state = {"messages": [HumanMessage(content=question)]}
|
| 131 |
+
result = self.graph.invoke(state, {'recursion_limit': 10})
|
| 132 |
messages = result.get("messages", [])
|
| 133 |
# Return only the content after "FINAL ANSWER:"
|
| 134 |
for msg in reversed(messages):
|