Update app.py
Browse files
app.py
CHANGED
|
@@ -10,7 +10,7 @@ import unicodedata
|
|
| 10 |
from typing import TypedDict, Annotated, Any, List, Optional
|
| 11 |
|
| 12 |
from huggingface_hub import InferenceClient
|
| 13 |
-
|
| 14 |
from langchain_core.messages import AnyMessage, HumanMessage, SystemMessage, AIMessage, ToolMessage
|
| 15 |
from langchain_core.tools import tool
|
| 16 |
from langchain_community.tools import DuckDuckGoSearchRun
|
|
@@ -148,11 +148,11 @@ class BasicAgent:
|
|
| 148 |
self.tools_by_name = {t.name: t for t in self.tools_list}
|
| 149 |
|
| 150 |
# InferenceClient diretto — usa la Serverless Inference API HF
|
| 151 |
-
|
| 152 |
-
api_key=os.getenv("
|
|
|
|
| 153 |
)
|
| 154 |
-
|
| 155 |
-
# Schema OpenAI dei tool per passarli al client
|
| 156 |
self.tools_schema = [_tool_to_openai_schema(t) for t in self.tools_list]
|
| 157 |
|
| 158 |
# Grafo LangGraph per gestire il loop ReAct
|
|
@@ -203,13 +203,9 @@ class BasicAgent:
|
|
| 203 |
sys_msg = SystemMessage(content=SYSTEM_PROMPT)
|
| 204 |
hf_messages = self._messages_to_hf_format([sys_msg] + state["messages"])
|
| 205 |
|
| 206 |
-
response = self.client.
|
| 207 |
-
model="
|
| 208 |
messages=hf_messages,
|
| 209 |
-
tools=self.tools_schema,
|
| 210 |
-
tool_choice="auto",
|
| 211 |
-
max_tokens=512,
|
| 212 |
-
temperature=0,
|
| 213 |
)
|
| 214 |
|
| 215 |
choice = response.choices[0].message
|
|
|
|
| 10 |
from typing import TypedDict, Annotated, Any, List, Optional
|
| 11 |
|
| 12 |
from huggingface_hub import InferenceClient
|
| 13 |
+
from openai import OpenAI
|
| 14 |
from langchain_core.messages import AnyMessage, HumanMessage, SystemMessage, AIMessage, ToolMessage
|
| 15 |
from langchain_core.tools import tool
|
| 16 |
from langchain_community.tools import DuckDuckGoSearchRun
|
|
|
|
| 148 |
self.tools_by_name = {t.name: t for t in self.tools_list}
|
| 149 |
|
| 150 |
# InferenceClient diretto — usa la Serverless Inference API HF
|
| 151 |
+
client = OpenAI(
|
| 152 |
+
api_key=os.getenv("GROQ_API_KEY"),
|
| 153 |
+
base_url="https://api.groq.com/openai/v1"
|
| 154 |
)
|
| 155 |
+
# Schema OpenAI dei tool per passarli al client
|
|
|
|
| 156 |
self.tools_schema = [_tool_to_openai_schema(t) for t in self.tools_list]
|
| 157 |
|
| 158 |
# Grafo LangGraph per gestire il loop ReAct
|
|
|
|
| 203 |
sys_msg = SystemMessage(content=SYSTEM_PROMPT)
|
| 204 |
hf_messages = self._messages_to_hf_format([sys_msg] + state["messages"])
|
| 205 |
|
| 206 |
+
response = self.client.chat.completions.create(
|
| 207 |
+
model="llama-3.3-70b-versatile",
|
| 208 |
messages=hf_messages,
|
|
|
|
|
|
|
|
|
|
|
|
|
| 209 |
)
|
| 210 |
|
| 211 |
choice = response.choices[0].message
|