mattibuzzo13 commited on
Commit
7605d87
·
verified ·
1 Parent(s): a884cd9

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +11 -7
app.py CHANGED
@@ -10,7 +10,7 @@ import unicodedata
10
  from typing import TypedDict, Annotated, Any, List, Optional
11
 
12
  from huggingface_hub import InferenceClient
13
- from openai import OpenAI
14
  from langchain_core.messages import AnyMessage, HumanMessage, SystemMessage, AIMessage, ToolMessage
15
  from langchain_core.tools import tool
16
  from langchain_community.tools import DuckDuckGoSearchRun
@@ -148,11 +148,11 @@ class BasicAgent:
148
  self.tools_by_name = {t.name: t for t in self.tools_list}
149
 
150
  # InferenceClient diretto — usa la Serverless Inference API HF
151
- client = OpenAI(
152
- api_key=os.getenv("GROQ_API_KEY"),
153
- base_url="https://api.groq.com/openai/v1"
154
  )
155
- # Schema OpenAI dei tool per passarli al client
 
156
  self.tools_schema = [_tool_to_openai_schema(t) for t in self.tools_list]
157
 
158
  # Grafo LangGraph per gestire il loop ReAct
@@ -203,9 +203,13 @@ class BasicAgent:
203
  sys_msg = SystemMessage(content=SYSTEM_PROMPT)
204
  hf_messages = self._messages_to_hf_format([sys_msg] + state["messages"])
205
 
206
- response = self.client.chat.completions.create(
207
- model="llama-3.3-70b-versatile",
208
  messages=hf_messages,
 
 
 
 
209
  )
210
 
211
  choice = response.choices[0].message
 
10
  from typing import TypedDict, Annotated, Any, List, Optional
11
 
12
  from huggingface_hub import InferenceClient
13
+
14
  from langchain_core.messages import AnyMessage, HumanMessage, SystemMessage, AIMessage, ToolMessage
15
  from langchain_core.tools import tool
16
  from langchain_community.tools import DuckDuckGoSearchRun
 
148
  self.tools_by_name = {t.name: t for t in self.tools_list}
149
 
150
  # InferenceClient diretto — usa la Serverless Inference API HF
151
+ self.client = InferenceClient(
152
+ api_key=os.getenv("HF_TOKEN"),
 
153
  )
154
+
155
+ # Schema OpenAI dei tool per passarli al client
156
  self.tools_schema = [_tool_to_openai_schema(t) for t in self.tools_list]
157
 
158
  # Grafo LangGraph per gestire il loop ReAct
 
203
  sys_msg = SystemMessage(content=SYSTEM_PROMPT)
204
  hf_messages = self._messages_to_hf_format([sys_msg] + state["messages"])
205
 
206
+ response = self.client.chat_completion(
207
+ model="Qwen/Qwen2.5-7B-Instruct",
208
  messages=hf_messages,
209
+ tools=self.tools_schema,
210
+ tool_choice="auto",
211
+ max_tokens=512,
212
+ temperature=0,
213
  )
214
 
215
  choice = response.choices[0].message