Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -21,7 +21,8 @@ pipe = DiffusionPipeline.from_pretrained(
|
|
| 21 |
attn_implementation="kernels-community/vllm-flash-attn3",
|
| 22 |
)
|
| 23 |
|
| 24 |
-
pipe.transformer.layers._repeated_blocks = ["ZImageTransformerBlock"]
|
|
|
|
| 25 |
|
| 26 |
@spaces.GPU
|
| 27 |
def generate_image(prompt, height, width, num_inference_steps, seed, randomize_seed):
|
|
@@ -38,6 +39,8 @@ def generate_image(prompt, height, width, num_inference_steps, seed, randomize_s
|
|
| 38 |
num_inference_steps=int(num_inference_steps),
|
| 39 |
guidance_scale=0.0, # Guidance should be 0 for Turbo models
|
| 40 |
generator=generator,
|
|
|
|
|
|
|
| 41 |
).images[0]
|
| 42 |
|
| 43 |
return image, seed
|
|
|
|
| 21 |
attn_implementation="kernels-community/vllm-flash-attn3",
|
| 22 |
)
|
| 23 |
|
| 24 |
+
#pipe.transformer.layers._repeated_blocks = ["ZImageTransformerBlock"]
|
| 25 |
+
#spaces.aoti_blocks_load(pipe.transformer.layers, "zerogpu-aoti/Z-Image", variant="fa3")
|
| 26 |
|
| 27 |
@spaces.GPU
|
| 28 |
def generate_image(prompt, height, width, num_inference_steps, seed, randomize_seed):
|
|
|
|
| 39 |
num_inference_steps=int(num_inference_steps),
|
| 40 |
guidance_scale=0.0, # Guidance should be 0 for Turbo models
|
| 41 |
generator=generator,
|
| 42 |
+
num_images_per_prompt=2
|
| 43 |
+
|
| 44 |
).images[0]
|
| 45 |
|
| 46 |
return image, seed
|