| | import asyncio |
| | import logging |
| | import os |
| | import random |
| | import tempfile |
| | import traceback |
| | import uuid |
| |
|
| | import aiohttp |
| | import torch |
| | from fastapi import FastAPI, HTTPException |
| | from fastapi.middleware.cors import CORSMiddleware |
| | from fastapi.staticfiles import StaticFiles |
| | from pydantic import BaseModel |
| |
|
| | from diffusers.pipelines.stable_diffusion_3 import StableDiffusion3Pipeline |
| |
|
| |
|
| | logger = logging.getLogger(__name__) |
| |
|
| |
|
| | class TextToImageInput(BaseModel): |
| | model: str |
| | prompt: str |
| | size: str | None = None |
| | n: int | None = None |
| |
|
| |
|
| | class HttpClient: |
| | session: aiohttp.ClientSession = None |
| |
|
| | def start(self): |
| | self.session = aiohttp.ClientSession() |
| |
|
| | async def stop(self): |
| | await self.session.close() |
| | self.session = None |
| |
|
| | def __call__(self) -> aiohttp.ClientSession: |
| | assert self.session is not None |
| | return self.session |
| |
|
| |
|
| | class TextToImagePipeline: |
| | pipeline: StableDiffusion3Pipeline = None |
| | device: str = None |
| |
|
| | def start(self): |
| | if torch.cuda.is_available(): |
| | model_path = os.getenv("MODEL_PATH", "stabilityai/stable-diffusion-3.5-large") |
| | logger.info("Loading CUDA") |
| | self.device = "cuda" |
| | self.pipeline = StableDiffusion3Pipeline.from_pretrained( |
| | model_path, |
| | torch_dtype=torch.bfloat16, |
| | ).to(device=self.device) |
| | elif torch.backends.mps.is_available(): |
| | model_path = os.getenv("MODEL_PATH", "stabilityai/stable-diffusion-3.5-medium") |
| | logger.info("Loading MPS for Mac M Series") |
| | self.device = "mps" |
| | self.pipeline = StableDiffusion3Pipeline.from_pretrained( |
| | model_path, |
| | torch_dtype=torch.bfloat16, |
| | ).to(device=self.device) |
| | else: |
| | raise Exception("No CUDA or MPS device available") |
| |
|
| |
|
| | app = FastAPI() |
| | service_url = os.getenv("SERVICE_URL", "http://localhost:8000") |
| | image_dir = os.path.join(tempfile.gettempdir(), "images") |
| | if not os.path.exists(image_dir): |
| | os.makedirs(image_dir) |
| | app.mount("/images", StaticFiles(directory=image_dir), name="images") |
| | http_client = HttpClient() |
| | shared_pipeline = TextToImagePipeline() |
| |
|
| | |
| | app.add_middleware( |
| | CORSMiddleware, |
| | allow_origins=["*"], |
| | allow_credentials=True, |
| | allow_methods=["*"], |
| | allow_headers=["*"], |
| | ) |
| |
|
| |
|
| | @app.on_event("startup") |
| | def startup(): |
| | http_client.start() |
| | shared_pipeline.start() |
| |
|
| |
|
| | def save_image(image): |
| | filename = "draw" + str(uuid.uuid4()).split("-")[0] + ".png" |
| | image_path = os.path.join(image_dir, filename) |
| | |
| | logger.info(f"Saving image to {image_path}") |
| | image.save(image_path) |
| | return os.path.join(service_url, "images", filename) |
| |
|
| |
|
| | @app.get("/") |
| | @app.post("/") |
| | @app.options("/") |
| | async def base(): |
| | return "Welcome to Diffusers! Where you can use diffusion models to generate images" |
| |
|
| |
|
| | @app.post("/v1/images/generations") |
| | async def generate_image(image_input: TextToImageInput): |
| | try: |
| | loop = asyncio.get_event_loop() |
| | scheduler = shared_pipeline.pipeline.scheduler.from_config(shared_pipeline.pipeline.scheduler.config) |
| | pipeline = StableDiffusion3Pipeline.from_pipe(shared_pipeline.pipeline, scheduler=scheduler) |
| | generator = torch.Generator(device=shared_pipeline.device) |
| | generator.manual_seed(random.randint(0, 10000000)) |
| | output = await loop.run_in_executor(None, lambda: pipeline(image_input.prompt, generator=generator)) |
| | logger.info(f"output: {output}") |
| | image_url = save_image(output.images[0]) |
| | return {"data": [{"url": image_url}]} |
| | except Exception as e: |
| | if isinstance(e, HTTPException): |
| | raise e |
| | elif hasattr(e, "message"): |
| | raise HTTPException(status_code=500, detail=e.message + traceback.format_exc()) |
| | raise HTTPException(status_code=500, detail=str(e) + traceback.format_exc()) |
| |
|
| |
|
| | if __name__ == "__main__": |
| | import uvicorn |
| |
|
| | uvicorn.run(app, host="0.0.0.0", port=8000) |
| |
|