Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -1,4 +1,4 @@
|
|
| 1 |
-
from smolagents import CodeAgent,
|
| 2 |
import datetime
|
| 3 |
import requests
|
| 4 |
import pytz
|
|
@@ -6,7 +6,7 @@ import yaml
|
|
| 6 |
import io
|
| 7 |
import base64
|
| 8 |
import schemdraw
|
| 9 |
-
|
| 10 |
from tools.final_answer import FinalAnswerTool
|
| 11 |
from Gradio_UI import GradioUI
|
| 12 |
|
|
@@ -37,14 +37,14 @@ def text_to_flowchart(steps_text: str) -> str:
|
|
| 37 |
if not parsed_steps:
|
| 38 |
parsed_steps = ["No steps provided."]
|
| 39 |
|
| 40 |
-
# Create the flowchart using SchemDraw
|
| 41 |
-
d = schemdraw.Drawing()
|
| 42 |
-
d.
|
| 43 |
for step in parsed_steps:
|
| 44 |
-
d.
|
| 45 |
-
d.
|
| 46 |
-
d.
|
| 47 |
-
d.
|
| 48 |
|
| 49 |
# Render diagram to PNG and encode as a data URL
|
| 50 |
buf = io.BytesIO()
|
|
@@ -76,18 +76,16 @@ def get_current_time_in_timezone(timezone: str) -> str:
|
|
| 76 |
|
| 77 |
final_answer = FinalAnswerTool()
|
| 78 |
|
| 79 |
-
model = HfApiModel(
|
| 80 |
max_tokens=2096,
|
| 81 |
temperature=0.5,
|
| 82 |
model_id='Qwen/Qwen2.5-Coder-32B-Instruct',
|
| 83 |
custom_role_conversions=None,
|
| 84 |
)
|
| 85 |
|
| 86 |
-
image_generation_tool = load_tool("agents-course/text-to-image", trust_remote_code=True)
|
| 87 |
-
|
| 88 |
with open("prompts.yaml", 'r') as stream:
|
| 89 |
prompt_templates = yaml.safe_load(stream)
|
| 90 |
-
|
| 91 |
agent = CodeAgent(
|
| 92 |
model=model,
|
| 93 |
tools=[text_to_flowchart, final_answer],
|
|
@@ -96,7 +94,7 @@ agent = CodeAgent(
|
|
| 96 |
grammar=None,
|
| 97 |
planning_interval=None,
|
| 98 |
name="Flowchart Generator",
|
| 99 |
-
description="Generates a flowchart diagram from
|
| 100 |
prompt_templates=prompt_templates
|
| 101 |
)
|
| 102 |
|
|
|
|
| 1 |
+
from smolagents import CodeAgent, tool, load_tool
|
| 2 |
import datetime
|
| 3 |
import requests
|
| 4 |
import pytz
|
|
|
|
| 6 |
import io
|
| 7 |
import base64
|
| 8 |
import schemdraw
|
| 9 |
+
from schemdraw import flow
|
| 10 |
from tools.final_answer import FinalAnswerTool
|
| 11 |
from Gradio_UI import GradioUI
|
| 12 |
|
|
|
|
| 37 |
if not parsed_steps:
|
| 38 |
parsed_steps = ["No steps provided."]
|
| 39 |
|
| 40 |
+
# Create the flowchart using SchemDraw's flow module
|
| 41 |
+
d = schemdraw.Drawing(unit=0.5, fontsize=10)
|
| 42 |
+
d += flow.Terminal().label("Start")
|
| 43 |
for step in parsed_steps:
|
| 44 |
+
d += flow.Arrow()
|
| 45 |
+
d += flow.Process().label(step)
|
| 46 |
+
d += flow.Arrow()
|
| 47 |
+
d += flow.Terminal().label("End")
|
| 48 |
|
| 49 |
# Render diagram to PNG and encode as a data URL
|
| 50 |
buf = io.BytesIO()
|
|
|
|
| 76 |
|
| 77 |
final_answer = FinalAnswerTool()
|
| 78 |
|
| 79 |
+
model = load_tool("HfApiModel", trust_remote_code=True)(
|
| 80 |
max_tokens=2096,
|
| 81 |
temperature=0.5,
|
| 82 |
model_id='Qwen/Qwen2.5-Coder-32B-Instruct',
|
| 83 |
custom_role_conversions=None,
|
| 84 |
)
|
| 85 |
|
|
|
|
|
|
|
| 86 |
with open("prompts.yaml", 'r') as stream:
|
| 87 |
prompt_templates = yaml.safe_load(stream)
|
| 88 |
+
|
| 89 |
agent = CodeAgent(
|
| 90 |
model=model,
|
| 91 |
tools=[text_to_flowchart, final_answer],
|
|
|
|
| 94 |
grammar=None,
|
| 95 |
planning_interval=None,
|
| 96 |
name="Flowchart Generator",
|
| 97 |
+
description="Generates a flowchart diagram from sequential steps using SchemDraw's flow module.",
|
| 98 |
prompt_templates=prompt_templates
|
| 99 |
)
|
| 100 |
|