File size: 4,633 Bytes
9b5b26a
 
 
 
c19d193
6aae614
9b5b26a
c5599c3
 
 
 
9b5b26a
5df72d6
c5599c3
b8a4aa3
c5599c3
 
b8a4aa3
c5599c3
b8a4aa3
c5599c3
 
 
 
b8a4aa3
 
 
 
 
 
 
 
c5599c3
 
 
 
b8a4aa3
 
 
 
 
 
 
 
c5599c3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9b5b26a
7e8503c
b8a4aa3
9b5b26a
7e8503c
9b5b26a
b8a4aa3
 
9b5b26a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8c01ffb
 
6aae614
ae7a494
 
3f91c02
ae7a494
e121372
bf6d34c
 
9e0fb74
 
fe328e0
13d500a
8c01ffb
 
9b5b26a
 
8c01ffb
861422e
 
9b5b26a
8c01ffb
8fe992b
c5599c3
8c01ffb
 
 
 
 
 
861422e
8fe992b
 
9b5b26a
8c01ffb
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
from smolagents import CodeAgent,DuckDuckGoSearchTool, HfApiModel,load_tool,tool
import datetime
import requests
import pytz
import yaml
from tools.final_answer import FinalAnswerTool
from Gradio_UI import GradioUI
from diffusers import StableDiffusionPipeline
import torch
from io import BytesIO
import base64

# Below is an example of a tool that does nothing. Amaze us with your creativity !

# @tool

class ImageGenerator:
    """A class to generate images from text prompts using Stable Diffusion."""
    def __init__(self, model_id="runwayml/stable-diffusion-v1-5", device="cuda" if torch.cuda.is_available() else "cpu"):
        """Initializes the ImageGenerator with the Stable Diffusion pipeline."""
        self.pipeline = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.float16 if device == "cuda" else torch.float32).to(device)
        self.device = device

    def generate_image(self, prompt, num_inference_steps=25, guidance_scale=7.5):
        """Generates an image from a text prompt.
        Args:
            prompt (str): The text prompt to generate the image from.
            num_inference_steps (int): The number of inference steps.
            guidance_scale (float): The guidance scale.
        Returns:
            PIL.Image.Image: The generated image.
        """
        image = self.pipeline(prompt, num_inference_steps=num_inference_steps, guidance_scale=guidance_scale).images[0]
        return image

    def generate_base64_image(self, prompt, num_inference_steps=25, guidance_scale=7.5):
        """Generates a base64 encoded image from a text prompt.
        Args:
            prompt (str): The text prompt to generate the image from.
            num_inference_steps (int): The number of inference steps.
            guidance_scale (float): The guidance scale.
        Returns:
            str: The base64 encoded image.
        """
        image = self.generate_image(prompt, num_inference_steps, guidance_scale)
        buffered = BytesIO()
        image.save(buffered, format="PNG")
        img_str = base64.b64encode(buffered.getvalue()).decode()
        return img_str

def generate_image_tool(image_generator):
    """Creates a tool function for image generation."""
    def image_generation_tool(prompt):
        """Generates an image from a prompt."""
        return image_generator.generate_base64_image(prompt)
    return image_generation_tool

# Initialize the ImageGenerator and tool
image_generator = ImageGenerator()
image_generation_tool_function = generate_image_tool(image_generator)

@tool
def generate_image_from_prompt(prompt: str) -> str:
    """Generates an image from a text prompt and embeds it in an HTML img tag.
    Args:
        prompt: The text prompt to generate the image from.
    """
    base64_image = image_generation_tool_function(prompt)
    return f'<img src="data:image/png;base64,{base64_image}" alt="Generated Image" />'

@tool
def get_current_time_in_timezone(timezone: str) -> str:
    """A tool that fetches the current local time in a specified timezone.
    Args:
        timezone: A string representing a valid timezone (e.g., 'America/New_York').
    """
    try:
        # Create timezone object
        tz = pytz.timezone(timezone)
        # Get current time in that timezone
        local_time = datetime.datetime.now(tz).strftime("%Y-%m-%d %H:%M:%S")
        return f"The current local time in {timezone} is: {local_time}"
    except Exception as e:
        return f"Error fetching time for timezone '{timezone}': {str(e)}"


final_answer = FinalAnswerTool()

# If the agent does not answer, the model is overloaded, please use another model or the following Hugging Face Endpoint that also contains qwen2.5 coder:
# model_id='https://pflgm2locj2t89co.us-east-1.aws.endpoints.huggingface.cloud' 

model = HfApiModel(
max_tokens=2096,
temperature=0.5,
model_id='https://pflgm2locj2t89co.us-east-1.aws.endpoints.huggingface.cloud',
# model_id='Qwen/Qwen2.5-Coder-32B-Instruct',# it is possible that this model may be overloaded
custom_role_conversions=None,
)


# Import tool from Hub
image_generation_tool = load_tool("agents-course/text-to-image", trust_remote_code=True)

with open("prompts.yaml", 'r') as stream:
    prompt_templates = yaml.safe_load(stream)
    
agent = CodeAgent(
    model=model,
    tools=[final_answer, generate_image_from_prompt, get_current_time_in_timezone], ## add your tools here (don't remove final answer)
    max_steps=6,
    verbosity_level=1,
    grammar=None,
    planning_interval=None,
    name=None,
    description=None,
    prompt_templates=prompt_templates
)


GradioUI(agent).launch()