Spaces:
Running on Zero
Running on Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -588,7 +588,8 @@ def sign_message(message, key):
|
|
| 588 |
signed_hash = base64.b64encode(hmac_digest).decode()
|
| 589 |
return signed_hash
|
| 590 |
|
| 591 |
-
|
|
|
|
| 592 |
if init_image is None:
|
| 593 |
do_img2img = False
|
| 594 |
|
|
@@ -597,11 +598,7 @@ def run(prompt, radio="model-v2", preset=PRESET_Q, h=1216, w=832, negative_promp
|
|
| 597 |
init_image = Image.fromarray(init_image)
|
| 598 |
init_image = init_image.resize((w, h))
|
| 599 |
init_image = np.array(init_image)
|
| 600 |
-
|
| 601 |
-
return zero_inference_api(prompt, radio, preset, h, w, negative_prompt, guidance_scale, randomize_seed, seed, do_img2img, init_image, image2image_strength, inference_steps=inference_steps)
|
| 602 |
-
|
| 603 |
-
@spaces.GPU
|
| 604 |
-
def zero_inference_api(prompt, radio="model-v2", preset=PRESET_Q, h=1216, w=832, negative_prompt=NEGATIVE_PROMPT, guidance_scale=4.0, randomize_seed=True, seed=42, do_img2img=False, init_image=None, image2image_strength=0, inference_steps=25, progress=gr.Progress(track_tqdm=True)):
|
| 605 |
prompt = prompt.strip() + ", " + preset.strip()
|
| 606 |
negative_prompt = negative_prompt.strip() if negative_prompt and negative_prompt.strip() else None
|
| 607 |
|
|
@@ -708,7 +705,6 @@ with gr.Blocks(theme=theme) as demo:
|
|
| 708 |
negative_prompt = gr.Textbox(label="Negative prompt", value=NEGATIVE_PROMPT, info="Is only applied for the CFG part, leave blank for unconditional generation")
|
| 709 |
seed = gr.Number(label="Seed", value=42, info="Seed for random number generator")
|
| 710 |
randomize_seed = gr.Checkbox(label="Randomize seed", value=True)
|
| 711 |
-
tpu_inference = gr.Checkbox(label="TPU Inference", value=False)
|
| 712 |
|
| 713 |
do_img2img = gr.Checkbox(label="Image to Image", value=False)
|
| 714 |
init_image = gr.Image(label="Input Image", visible=False)
|
|
@@ -731,7 +727,7 @@ with gr.Blocks(theme=theme) as demo:
|
|
| 731 |
prompt.submit
|
| 732 |
],
|
| 733 |
fn=run,
|
| 734 |
-
inputs=[prompt, radio, preset, height, width, negative_prompt, guidance_scale, randomize_seed, seed,
|
| 735 |
outputs=[output, seed],
|
| 736 |
concurrency_limit=1,
|
| 737 |
)
|
|
|
|
| 588 |
signed_hash = base64.b64encode(hmac_digest).decode()
|
| 589 |
return signed_hash
|
| 590 |
|
| 591 |
+
@spaces.GPU
|
| 592 |
+
def run(prompt, radio="model-v2", preset=PRESET_Q, h=1216, w=832, negative_prompt=NEGATIVE_PROMPT, guidance_scale=4.0, randomize_seed=True, seed=42, do_img2img=False, init_image=None, image2image_strength=0, inference_steps=25, progress=gr.Progress(track_tqdm=True)):
|
| 593 |
if init_image is None:
|
| 594 |
do_img2img = False
|
| 595 |
|
|
|
|
| 598 |
init_image = Image.fromarray(init_image)
|
| 599 |
init_image = init_image.resize((w, h))
|
| 600 |
init_image = np.array(init_image)
|
| 601 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
| 602 |
prompt = prompt.strip() + ", " + preset.strip()
|
| 603 |
negative_prompt = negative_prompt.strip() if negative_prompt and negative_prompt.strip() else None
|
| 604 |
|
|
|
|
| 705 |
negative_prompt = gr.Textbox(label="Negative prompt", value=NEGATIVE_PROMPT, info="Is only applied for the CFG part, leave blank for unconditional generation")
|
| 706 |
seed = gr.Number(label="Seed", value=42, info="Seed for random number generator")
|
| 707 |
randomize_seed = gr.Checkbox(label="Randomize seed", value=True)
|
|
|
|
| 708 |
|
| 709 |
do_img2img = gr.Checkbox(label="Image to Image", value=False)
|
| 710 |
init_image = gr.Image(label="Input Image", visible=False)
|
|
|
|
| 727 |
prompt.submit
|
| 728 |
],
|
| 729 |
fn=run,
|
| 730 |
+
inputs=[prompt, radio, preset, height, width, negative_prompt, guidance_scale, randomize_seed, seed, do_img2img, init_image, image2image_resize, image2image_strength, inference_steps],
|
| 731 |
outputs=[output, seed],
|
| 732 |
concurrency_limit=1,
|
| 733 |
)
|