nyanko7 commited on
Commit
987afc5
·
verified ·
1 Parent(s): 4cffd8a

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +4 -8
app.py CHANGED
@@ -588,7 +588,8 @@ def sign_message(message, key):
588
  signed_hash = base64.b64encode(hmac_digest).decode()
589
  return signed_hash
590
 
591
- def run(prompt, radio="model-v2", preset=PRESET_Q, h=1216, w=832, negative_prompt=NEGATIVE_PROMPT, guidance_scale=4.0, randomize_seed=True, seed=42, tpu_inference=False, do_img2img=False, init_image=None, image2image_resize=False, image2image_strength=0, inference_steps=25, progress=gr.Progress(track_tqdm=True)):
 
592
  if init_image is None:
593
  do_img2img = False
594
 
@@ -597,11 +598,7 @@ def run(prompt, radio="model-v2", preset=PRESET_Q, h=1216, w=832, negative_promp
597
  init_image = Image.fromarray(init_image)
598
  init_image = init_image.resize((w, h))
599
  init_image = np.array(init_image)
600
-
601
- return zero_inference_api(prompt, radio, preset, h, w, negative_prompt, guidance_scale, randomize_seed, seed, do_img2img, init_image, image2image_strength, inference_steps=inference_steps)
602
-
603
- @spaces.GPU
604
- def zero_inference_api(prompt, radio="model-v2", preset=PRESET_Q, h=1216, w=832, negative_prompt=NEGATIVE_PROMPT, guidance_scale=4.0, randomize_seed=True, seed=42, do_img2img=False, init_image=None, image2image_strength=0, inference_steps=25, progress=gr.Progress(track_tqdm=True)):
605
  prompt = prompt.strip() + ", " + preset.strip()
606
  negative_prompt = negative_prompt.strip() if negative_prompt and negative_prompt.strip() else None
607
 
@@ -708,7 +705,6 @@ with gr.Blocks(theme=theme) as demo:
708
  negative_prompt = gr.Textbox(label="Negative prompt", value=NEGATIVE_PROMPT, info="Is only applied for the CFG part, leave blank for unconditional generation")
709
  seed = gr.Number(label="Seed", value=42, info="Seed for random number generator")
710
  randomize_seed = gr.Checkbox(label="Randomize seed", value=True)
711
- tpu_inference = gr.Checkbox(label="TPU Inference", value=False)
712
 
713
  do_img2img = gr.Checkbox(label="Image to Image", value=False)
714
  init_image = gr.Image(label="Input Image", visible=False)
@@ -731,7 +727,7 @@ with gr.Blocks(theme=theme) as demo:
731
  prompt.submit
732
  ],
733
  fn=run,
734
- inputs=[prompt, radio, preset, height, width, negative_prompt, guidance_scale, randomize_seed, seed, tpu_inference, do_img2img, init_image, image2image_resize, image2image_strength, inference_steps],
735
  outputs=[output, seed],
736
  concurrency_limit=1,
737
  )
 
588
  signed_hash = base64.b64encode(hmac_digest).decode()
589
  return signed_hash
590
 
591
+ @spaces.GPU
592
+ def run(prompt, radio="model-v2", preset=PRESET_Q, h=1216, w=832, negative_prompt=NEGATIVE_PROMPT, guidance_scale=4.0, randomize_seed=True, seed=42, do_img2img=False, init_image=None, image2image_strength=0, inference_steps=25, progress=gr.Progress(track_tqdm=True)):
593
  if init_image is None:
594
  do_img2img = False
595
 
 
598
  init_image = Image.fromarray(init_image)
599
  init_image = init_image.resize((w, h))
600
  init_image = np.array(init_image)
601
+
 
 
 
 
602
  prompt = prompt.strip() + ", " + preset.strip()
603
  negative_prompt = negative_prompt.strip() if negative_prompt and negative_prompt.strip() else None
604
 
 
705
  negative_prompt = gr.Textbox(label="Negative prompt", value=NEGATIVE_PROMPT, info="Is only applied for the CFG part, leave blank for unconditional generation")
706
  seed = gr.Number(label="Seed", value=42, info="Seed for random number generator")
707
  randomize_seed = gr.Checkbox(label="Randomize seed", value=True)
 
708
 
709
  do_img2img = gr.Checkbox(label="Image to Image", value=False)
710
  init_image = gr.Image(label="Input Image", visible=False)
 
727
  prompt.submit
728
  ],
729
  fn=run,
730
+ inputs=[prompt, radio, preset, height, width, negative_prompt, guidance_scale, randomize_seed, seed, do_img2img, init_image, image2image_resize, image2image_strength, inference_steps],
731
  outputs=[output, seed],
732
  concurrency_limit=1,
733
  )