Thanh Vinh Vo
commited on
Commit
·
58c410b
1
Parent(s):
9b9abfe
update
Browse files
app.py
CHANGED
|
@@ -177,7 +177,7 @@ class BasicAgent:
|
|
| 177 |
print("BasicAgent initialized.")
|
| 178 |
self.multimodal_agent = CodeAgent(
|
| 179 |
tools=[VisitWebpageTool(), GoogleSearchTool("serper"), get_file, audio_to_text],
|
| 180 |
-
model= OpenAIServerModel(model_id="gpt-4o"),
|
| 181 |
additional_authorized_imports=[
|
| 182 |
"requests",
|
| 183 |
"bs4",
|
|
@@ -200,13 +200,13 @@ class BasicAgent:
|
|
| 200 |
This agent can reason across audio, vision, and text, a.k.a multimodal agent. """,
|
| 201 |
verbosity_level=0,
|
| 202 |
max_steps=10,
|
| 203 |
-
temperature=0.0,
|
| 204 |
)
|
| 205 |
|
| 206 |
self.code_agent = CodeAgent(
|
| 207 |
tools=[VisitWebpageTool(), GoogleSearchTool("serper"), get_file, audio_to_text, extract_table_from_html, WikipediaSearchTool()],
|
| 208 |
model=InferenceClientModel(
|
| 209 |
model_id="Qwen/Qwen2.5-Coder-32B-Instruct",
|
|
|
|
| 210 |
),
|
| 211 |
additional_authorized_imports=[
|
| 212 |
"requests",
|
|
@@ -243,12 +243,12 @@ class BasicAgent:
|
|
| 243 |
""",
|
| 244 |
verbosity_level=0,
|
| 245 |
max_steps=10,
|
| 246 |
-
temperature=0.0,
|
| 247 |
)
|
| 248 |
|
| 249 |
self.manager_agent = CodeAgent(
|
| 250 |
model=InferenceClientModel(
|
| 251 |
-
"Qwen/Qwen2.5-32B-Instruct"
|
|
|
|
| 252 |
),
|
| 253 |
tools=[get_file, audio_to_text, extract_table_from_html],
|
| 254 |
managed_agents=[
|
|
@@ -277,7 +277,6 @@ class BasicAgent:
|
|
| 277 |
],
|
| 278 |
planning_interval=5,
|
| 279 |
max_steps=15,
|
| 280 |
-
temperature=0.0,
|
| 281 |
)
|
| 282 |
|
| 283 |
def __call__(self, question: str, question_id: str, file_name: str) -> str:
|
|
|
|
| 177 |
print("BasicAgent initialized.")
|
| 178 |
self.multimodal_agent = CodeAgent(
|
| 179 |
tools=[VisitWebpageTool(), GoogleSearchTool("serper"), get_file, audio_to_text],
|
| 180 |
+
model= OpenAIServerModel(model_id="gpt-4o", temperature=0.0,),
|
| 181 |
additional_authorized_imports=[
|
| 182 |
"requests",
|
| 183 |
"bs4",
|
|
|
|
| 200 |
This agent can reason across audio, vision, and text, a.k.a multimodal agent. """,
|
| 201 |
verbosity_level=0,
|
| 202 |
max_steps=10,
|
|
|
|
| 203 |
)
|
| 204 |
|
| 205 |
self.code_agent = CodeAgent(
|
| 206 |
tools=[VisitWebpageTool(), GoogleSearchTool("serper"), get_file, audio_to_text, extract_table_from_html, WikipediaSearchTool()],
|
| 207 |
model=InferenceClientModel(
|
| 208 |
model_id="Qwen/Qwen2.5-Coder-32B-Instruct",
|
| 209 |
+
temperature=0.0,
|
| 210 |
),
|
| 211 |
additional_authorized_imports=[
|
| 212 |
"requests",
|
|
|
|
| 243 |
""",
|
| 244 |
verbosity_level=0,
|
| 245 |
max_steps=10,
|
|
|
|
| 246 |
)
|
| 247 |
|
| 248 |
self.manager_agent = CodeAgent(
|
| 249 |
model=InferenceClientModel(
|
| 250 |
+
model_id="Qwen/Qwen2.5-32B-Instruct",
|
| 251 |
+
temperature=0.0,
|
| 252 |
),
|
| 253 |
tools=[get_file, audio_to_text, extract_table_from_html],
|
| 254 |
managed_agents=[
|
|
|
|
| 277 |
],
|
| 278 |
planning_interval=5,
|
| 279 |
max_steps=15,
|
|
|
|
| 280 |
)
|
| 281 |
|
| 282 |
def __call__(self, question: str, question_id: str, file_name: str) -> str:
|