Spaces:
Running
on
T4
Running
on
T4
Update app.py
Browse files
app.py
CHANGED
|
@@ -15,7 +15,7 @@ def genie (Prompt, scale, steps, Seed):
|
|
| 15 |
generator = torch.Generator(device=device).manual_seed(Seed)
|
| 16 |
#images = pipe(prompt, num_inference_steps=steps, guidance_scale=scale, generator=generator).images[0]
|
| 17 |
low_res_latents = pipe(Prompt, num_inference_steps=steps, guidance_scale=scale, generator=generator, output_type="latent").images
|
| 18 |
-
upscaled_image = upscaler(prompt='', image=low_res_latents, num_inference_steps=
|
| 19 |
return upscaled_image
|
| 20 |
|
| 21 |
gr.Interface(fn=genie, inputs=[gr.Textbox(label='What you want the AI to generate. 77 Token Limit.'),
|
|
|
|
| 15 |
generator = torch.Generator(device=device).manual_seed(Seed)
|
| 16 |
#images = pipe(prompt, num_inference_steps=steps, guidance_scale=scale, generator=generator).images[0]
|
| 17 |
low_res_latents = pipe(Prompt, num_inference_steps=steps, guidance_scale=scale, generator=generator, output_type="latent").images
|
| 18 |
+
upscaled_image = upscaler(prompt='', image=low_res_latents, num_inference_steps=5, guidance_scale=0, generator=generator).images[0]
|
| 19 |
return upscaled_image
|
| 20 |
|
| 21 |
gr.Interface(fn=genie, inputs=[gr.Textbox(label='What you want the AI to generate. 77 Token Limit.'),
|