whiran commited on
Commit
4aff01c
·
verified ·
1 Parent(s): 053e4b6

Update app.py

Browse files

Add new diffusion model as the image gen

Files changed (1) hide show
  1. app.py +49 -31
app.py CHANGED
@@ -1,69 +1,87 @@
1
- from smolagents import CodeAgent,DuckDuckGoSearchTool, HfApiModel,load_tool,tool
2
  import datetime
3
  import requests
4
  import pytz
5
  import yaml
 
 
 
 
6
  from tools.final_answer import FinalAnswerTool
7
-
8
  from Gradio_UI import GradioUI
9
 
10
- # Below is an example of a tool that does nothing. Amaze us with your creativity !
11
  @tool
12
- def my_custom_tool(arg1:str, arg2:int)-> str: #it's import to specify the return type
13
- #Keep this format for the description / args / args description but feel free to modify the tool
14
- """A tool that does nothing yet
 
 
 
 
15
  Args:
16
- arg1: the first argument
17
- arg2: the second argument
 
 
18
  """
19
- return "What magic will you build ?"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
 
21
  @tool
22
  def get_current_time_in_timezone(timezone: str) -> str:
23
  """A tool that fetches the current local time in a specified timezone.
24
  Args:
25
- timezone: A string representing a valid timezone (e.g., 'America/New_York').
26
  """
27
  try:
28
- # Create timezone object
29
  tz = pytz.timezone(timezone)
30
- # Get current time in that timezone
31
  local_time = datetime.datetime.now(tz).strftime("%Y-%m-%d %H:%M:%S")
32
- return f"The current local time in {timezone} is: {local_time}"
33
  except Exception as e:
34
- return f"Error fetching time for timezone '{timezone}': {str(e)}"
35
-
36
 
37
  final_answer = FinalAnswerTool()
38
 
39
- # If the agent does not answer, the model is overloaded, please use another model or the following Hugging Face Endpoint that also contains qwen2.5 coder:
40
- # model_id='https://pflgm2locj2t89co.us-east-1.aws.endpoints.huggingface.cloud'
41
-
42
  model = HfApiModel(
43
- max_tokens=2096,
44
- temperature=0.5,
45
- model_id='Qwen/Qwen2.5-Coder-32B-Instruct',# it is possible that this model may be overloaded
46
- custom_role_conversions=None,
47
  )
48
 
49
-
50
- # Import tool from Hub
51
- image_generation_tool = load_tool("agents-course/text-to-image", trust_remote_code=True)
52
-
53
  with open("prompts.yaml", 'r') as stream:
54
  prompt_templates = yaml.safe_load(stream)
55
 
56
  agent = CodeAgent(
57
  model=model,
58
- tools=[image_generation_tool,final_answer], ## add your tools here (don't remove final answer)
59
  max_steps=6,
60
  verbosity_level=1,
 
61
  grammar=None,
62
  planning_interval=None,
63
- name=None,
64
- description=None,
65
- prompt_templates=prompt_templates
66
  )
67
 
68
-
69
  GradioUI(agent).launch()
 
1
+ from smolagents import CodeAgent, DuckDuckGoSearchTool, HfApiModel, load_tool, tool
2
  import datetime
3
  import requests
4
  import pytz
5
  import yaml
6
+ import torch
7
+ from diffusers import StableDiffusionPipeline
8
+ from pathlib import Path
9
+ from typing import Optional
10
  from tools.final_answer import FinalAnswerTool
 
11
  from Gradio_UI import GradioUI
12
 
 
13
  @tool
14
+ def generate_image_tool(
15
+ prompt: str,
16
+ model_name: str = "stabilityai/stable-diffusion-2-1",
17
+ num_inference_steps: int = 50,
18
+ output_path: str = "generated_image.png"
19
+ ) -> str:
20
+ """A tool for generating images from text prompts using Stable Diffusion.
21
  Args:
22
+ prompt: Text description of the image (must be SFW)
23
+ model_name: AI model to use (default: stabilityai/stable-diffusion-2-1)
24
+ num_inference_steps: Quality steps (20-100)
25
+ output_path: Where to save the image
26
  """
27
+ # Safety check
28
+ unsafe_keywords = ["nude", "porn", "explicit", "adult", "nsfw"]
29
+ if any(kw in prompt.lower() for kw in unsafe_keywords):
30
+ return "Error: Content policy violation detected in prompt"
31
+
32
+ try:
33
+ pipe = StableDiffusionPipeline.from_pretrained(
34
+ model_name,
35
+ torch_dtype=torch.float16
36
+ ).to("cuda")
37
+
38
+ image = pipe(
39
+ prompt=prompt,
40
+ num_inference_steps=num_inference_steps
41
+ ).images[0]
42
+
43
+ Path(output_path).parent.mkdir(parents=True, exist_ok=True)
44
+ image.save(output_path)
45
+ return f"Image generated successfully at: {output_path}"
46
+
47
+ except Exception as e:
48
+ return f"Image generation failed: {str(e)}"
49
 
50
  @tool
51
  def get_current_time_in_timezone(timezone: str) -> str:
52
  """A tool that fetches the current local time in a specified timezone.
53
  Args:
54
+ timezone: A valid timezone (e.g., 'America/New_York')
55
  """
56
  try:
 
57
  tz = pytz.timezone(timezone)
 
58
  local_time = datetime.datetime.now(tz).strftime("%Y-%m-%d %H:%M:%S")
59
+ return f"Current time in {timezone}: {local_time}"
60
  except Exception as e:
61
+ return f"Error: {str(e)}"
 
62
 
63
  final_answer = FinalAnswerTool()
64
 
 
 
 
65
  model = HfApiModel(
66
+ max_tokens=2096,
67
+ temperature=0.5,
68
+ model_id='Qwen/Qwen2.5-Coder-32B-Instruct',
69
+ custom_role_conversions=None,
70
  )
71
 
 
 
 
 
72
  with open("prompts.yaml", 'r') as stream:
73
  prompt_templates = yaml.safe_load(stream)
74
 
75
  agent = CodeAgent(
76
  model=model,
77
+ tools=[generate_image_tool, get_current_time_in_timezone, final_answer],
78
  max_steps=6,
79
  verbosity_level=1,
80
+ prompt_templates=prompt_templates,
81
  grammar=None,
82
  planning_interval=None,
83
+ name="Creative Assistant",
84
+ description="AI assistant capable of generating images and providing time information"
 
85
  )
86
 
 
87
  GradioUI(agent).launch()