Spaces:
Sleeping
Sleeping
File size: 4,633 Bytes
9b5b26a c19d193 6aae614 9b5b26a c5599c3 9b5b26a 5df72d6 c5599c3 b8a4aa3 c5599c3 b8a4aa3 c5599c3 b8a4aa3 c5599c3 b8a4aa3 c5599c3 b8a4aa3 c5599c3 9b5b26a 7e8503c b8a4aa3 9b5b26a 7e8503c 9b5b26a b8a4aa3 9b5b26a 8c01ffb 6aae614 ae7a494 3f91c02 ae7a494 e121372 bf6d34c 9e0fb74 fe328e0 13d500a 8c01ffb 9b5b26a 8c01ffb 861422e 9b5b26a 8c01ffb 8fe992b c5599c3 8c01ffb 861422e 8fe992b 9b5b26a 8c01ffb |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 |
from smolagents import CodeAgent,DuckDuckGoSearchTool, HfApiModel,load_tool,tool
import datetime
import requests
import pytz
import yaml
from tools.final_answer import FinalAnswerTool
from Gradio_UI import GradioUI
from diffusers import StableDiffusionPipeline
import torch
from io import BytesIO
import base64
# Below is an example of a tool that does nothing. Amaze us with your creativity !
# @tool
class ImageGenerator:
"""A class to generate images from text prompts using Stable Diffusion."""
def __init__(self, model_id="runwayml/stable-diffusion-v1-5", device="cuda" if torch.cuda.is_available() else "cpu"):
"""Initializes the ImageGenerator with the Stable Diffusion pipeline."""
self.pipeline = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.float16 if device == "cuda" else torch.float32).to(device)
self.device = device
def generate_image(self, prompt, num_inference_steps=25, guidance_scale=7.5):
"""Generates an image from a text prompt.
Args:
prompt (str): The text prompt to generate the image from.
num_inference_steps (int): The number of inference steps.
guidance_scale (float): The guidance scale.
Returns:
PIL.Image.Image: The generated image.
"""
image = self.pipeline(prompt, num_inference_steps=num_inference_steps, guidance_scale=guidance_scale).images[0]
return image
def generate_base64_image(self, prompt, num_inference_steps=25, guidance_scale=7.5):
"""Generates a base64 encoded image from a text prompt.
Args:
prompt (str): The text prompt to generate the image from.
num_inference_steps (int): The number of inference steps.
guidance_scale (float): The guidance scale.
Returns:
str: The base64 encoded image.
"""
image = self.generate_image(prompt, num_inference_steps, guidance_scale)
buffered = BytesIO()
image.save(buffered, format="PNG")
img_str = base64.b64encode(buffered.getvalue()).decode()
return img_str
def generate_image_tool(image_generator):
"""Creates a tool function for image generation."""
def image_generation_tool(prompt):
"""Generates an image from a prompt."""
return image_generator.generate_base64_image(prompt)
return image_generation_tool
# Initialize the ImageGenerator and tool
image_generator = ImageGenerator()
image_generation_tool_function = generate_image_tool(image_generator)
@tool
def generate_image_from_prompt(prompt: str) -> str:
"""Generates an image from a text prompt and embeds it in an HTML img tag.
Args:
prompt: The text prompt to generate the image from.
"""
base64_image = image_generation_tool_function(prompt)
return f'<img src="data:image/png;base64,{base64_image}" alt="Generated Image" />'
@tool
def get_current_time_in_timezone(timezone: str) -> str:
"""A tool that fetches the current local time in a specified timezone.
Args:
timezone: A string representing a valid timezone (e.g., 'America/New_York').
"""
try:
# Create timezone object
tz = pytz.timezone(timezone)
# Get current time in that timezone
local_time = datetime.datetime.now(tz).strftime("%Y-%m-%d %H:%M:%S")
return f"The current local time in {timezone} is: {local_time}"
except Exception as e:
return f"Error fetching time for timezone '{timezone}': {str(e)}"
final_answer = FinalAnswerTool()
# If the agent does not answer, the model is overloaded, please use another model or the following Hugging Face Endpoint that also contains qwen2.5 coder:
# model_id='https://pflgm2locj2t89co.us-east-1.aws.endpoints.huggingface.cloud'
model = HfApiModel(
max_tokens=2096,
temperature=0.5,
model_id='https://pflgm2locj2t89co.us-east-1.aws.endpoints.huggingface.cloud',
# model_id='Qwen/Qwen2.5-Coder-32B-Instruct',# it is possible that this model may be overloaded
custom_role_conversions=None,
)
# Import tool from Hub
image_generation_tool = load_tool("agents-course/text-to-image", trust_remote_code=True)
with open("prompts.yaml", 'r') as stream:
prompt_templates = yaml.safe_load(stream)
agent = CodeAgent(
model=model,
tools=[final_answer, generate_image_from_prompt, get_current_time_in_timezone], ## add your tools here (don't remove final answer)
max_steps=6,
verbosity_level=1,
grammar=None,
planning_interval=None,
name=None,
description=None,
prompt_templates=prompt_templates
)
GradioUI(agent).launch() |