Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -2,6 +2,8 @@ import torch
|
|
| 2 |
import spaces
|
| 3 |
import gradio as gr
|
| 4 |
from diffusers import ZImagePipeline, ZImageTransformer2DModel
|
|
|
|
|
|
|
| 5 |
|
| 6 |
# Load the pipeline once at startup
|
| 7 |
print("Loading Z-Image-Turbo pipeline...")
|
|
@@ -23,12 +25,15 @@ pipe.to("cuda")
|
|
| 23 |
# spaces.aoti_blocks_load(pipe.transformer.layers, "zerogpu-aoti/Z-Image", variant="fa3")
|
| 24 |
|
| 25 |
print("Pipeline loaded!")
|
|
|
|
|
|
|
|
|
|
| 26 |
|
| 27 |
@spaces.GPU
|
| 28 |
-
def generate_image(prompt, height, width, num_inference_steps, seed, randomize_seed, progress=gr.Progress(track_tqdm=True)):
|
| 29 |
"""Generate an image from the given prompt."""
|
| 30 |
if randomize_seed:
|
| 31 |
-
seed =
|
| 32 |
|
| 33 |
generator = torch.Generator("cuda").manual_seed(int(seed))
|
| 34 |
image = pipe(
|
|
|
|
| 2 |
import spaces
|
| 3 |
import gradio as gr
|
| 4 |
from diffusers import ZImagePipeline, ZImageTransformer2DModel
|
| 5 |
+
import numpy as np
|
| 6 |
+
import random
|
| 7 |
|
| 8 |
# Load the pipeline once at startup
|
| 9 |
print("Loading Z-Image-Turbo pipeline...")
|
|
|
|
| 25 |
# spaces.aoti_blocks_load(pipe.transformer.layers, "zerogpu-aoti/Z-Image", variant="fa3")
|
| 26 |
|
| 27 |
print("Pipeline loaded!")
|
| 28 |
+
MAX_SEED = np.iinfo(np.int32).max
|
| 29 |
+
|
| 30 |
+
|
| 31 |
|
| 32 |
@spaces.GPU
|
| 33 |
+
def generate_image(prompt, height, width, num_inference_steps=10, seed=42, randomize_seed=True, progress=gr.Progress(track_tqdm=True)):
|
| 34 |
"""Generate an image from the given prompt."""
|
| 35 |
if randomize_seed:
|
| 36 |
+
seed = random.randint(0, MAX_SEED)
|
| 37 |
|
| 38 |
generator = torch.Generator("cuda").manual_seed(int(seed))
|
| 39 |
image = pipe(
|