Update app.py
Browse files
app.py
CHANGED
|
@@ -10,7 +10,7 @@ import unicodedata
|
|
| 10 |
from typing import TypedDict, Annotated, Any, List, Optional
|
| 11 |
|
| 12 |
from huggingface_hub import InferenceClient
|
| 13 |
-
|
| 14 |
from langchain_core.messages import AnyMessage, HumanMessage, SystemMessage, AIMessage, ToolMessage
|
| 15 |
from langchain_core.tools import tool
|
| 16 |
from langchain_community.tools import DuckDuckGoSearchRun
|
|
@@ -148,11 +148,11 @@ class BasicAgent:
|
|
| 148 |
self.tools_by_name = {t.name: t for t in self.tools_list}
|
| 149 |
|
| 150 |
# InferenceClient diretto — usa la Serverless Inference API HF
|
| 151 |
-
client =
|
| 152 |
-
api_key=os.getenv("
|
| 153 |
-
base_url="https://api.groq.com/openai/v1"
|
| 154 |
)
|
| 155 |
-
|
|
|
|
| 156 |
self.tools_schema = [_tool_to_openai_schema(t) for t in self.tools_list]
|
| 157 |
|
| 158 |
# Grafo LangGraph per gestire il loop ReAct
|
|
@@ -203,9 +203,13 @@ class BasicAgent:
|
|
| 203 |
sys_msg = SystemMessage(content=SYSTEM_PROMPT)
|
| 204 |
hf_messages = self._messages_to_hf_format([sys_msg] + state["messages"])
|
| 205 |
|
| 206 |
-
response = self.client.
|
| 207 |
-
model="
|
| 208 |
messages=hf_messages,
|
|
|
|
|
|
|
|
|
|
|
|
|
| 209 |
)
|
| 210 |
|
| 211 |
choice = response.choices[0].message
|
|
|
|
| 10 |
from typing import TypedDict, Annotated, Any, List, Optional
|
| 11 |
|
| 12 |
from huggingface_hub import InferenceClient
|
| 13 |
+
|
| 14 |
from langchain_core.messages import AnyMessage, HumanMessage, SystemMessage, AIMessage, ToolMessage
|
| 15 |
from langchain_core.tools import tool
|
| 16 |
from langchain_community.tools import DuckDuckGoSearchRun
|
|
|
|
| 148 |
self.tools_by_name = {t.name: t for t in self.tools_list}
|
| 149 |
|
| 150 |
# InferenceClient diretto — usa la Serverless Inference API HF
|
| 151 |
+
self.client = InferenceClient(
|
| 152 |
+
api_key=os.getenv("HF_TOKEN"),
|
|
|
|
| 153 |
)
|
| 154 |
+
|
| 155 |
+
# Schema OpenAI dei tool per passarli al client
|
| 156 |
self.tools_schema = [_tool_to_openai_schema(t) for t in self.tools_list]
|
| 157 |
|
| 158 |
# Grafo LangGraph per gestire il loop ReAct
|
|
|
|
| 203 |
sys_msg = SystemMessage(content=SYSTEM_PROMPT)
|
| 204 |
hf_messages = self._messages_to_hf_format([sys_msg] + state["messages"])
|
| 205 |
|
| 206 |
+
response = self.client.chat_completion(
|
| 207 |
+
model="Qwen/Qwen2.5-7B-Instruct",
|
| 208 |
messages=hf_messages,
|
| 209 |
+
tools=self.tools_schema,
|
| 210 |
+
tool_choice="auto",
|
| 211 |
+
max_tokens=512,
|
| 212 |
+
temperature=0,
|
| 213 |
)
|
| 214 |
|
| 215 |
choice = response.choices[0].message
|