Commit
·
7e5566d
1
Parent(s):
744a1fc
Capitalise button
Browse files
app.py
CHANGED
|
@@ -89,14 +89,14 @@ with gr.Blocks() as mindseye:
|
|
| 89 |
style = gr.inputs.Dropdown(label="Style - Hyper Fast Results is fast but compromises a bit of the quality",choices=["Default","Balanced","Detailed","Consistent Creativity","Realistic","Smooth","Subtle MSE","Hyper Fast Results"],default="Hyper Fast Results")
|
| 90 |
steps = gr.inputs.Slider(label="Steps - more steps can increase quality but will take longer to generate. All styles that are not Hyper Fast need at least 200 steps",default=50,maximum=300,minimum=1,step=1)
|
| 91 |
flavor = gr.inputs.Dropdown(label="Flavor - pick a flavor for the style of the images, based on the images below",choices=["ginger", "cumin", "holywater", "zynth", "wyvern", "aaron", "moth", "juu"])
|
| 92 |
-
get_image_vqgan = gr.
|
| 93 |
with gr.TabItem("Guided Diffusion"):
|
| 94 |
gr.Markdown("Guided Diffusion models produce superb quality results. V-Diffusion is its latest implementation")
|
| 95 |
steps_diff = gr.inputs.Slider(label="Steps - more steps can increase quality but will take longer to generate",default=40,maximum=80,minimum=1,step=1)
|
| 96 |
images_diff = gr.inputs.Slider(label="Number of images in parallel", default=2, maximum=4, minimum=1, step=1)
|
| 97 |
weight = gr.inputs.Slider(label="Weight - how closely the image should resemble the prompt", default=5, maximum=15, minimum=0, step=1)
|
| 98 |
clip = gr.inputs.Checkbox(label="CLIP Guided - improves coherence with complex prompts, makes it slower")
|
| 99 |
-
get_image_diffusion = gr.
|
| 100 |
with gr.Row():
|
| 101 |
with gr.Tabs():
|
| 102 |
#with gr.TabItem("Image output"):
|
|
|
|
| 89 |
style = gr.inputs.Dropdown(label="Style - Hyper Fast Results is fast but compromises a bit of the quality",choices=["Default","Balanced","Detailed","Consistent Creativity","Realistic","Smooth","Subtle MSE","Hyper Fast Results"],default="Hyper Fast Results")
|
| 90 |
steps = gr.inputs.Slider(label="Steps - more steps can increase quality but will take longer to generate. All styles that are not Hyper Fast need at least 200 steps",default=50,maximum=300,minimum=1,step=1)
|
| 91 |
flavor = gr.inputs.Dropdown(label="Flavor - pick a flavor for the style of the images, based on the images below",choices=["ginger", "cumin", "holywater", "zynth", "wyvern", "aaron", "moth", "juu"])
|
| 92 |
+
get_image_vqgan = gr.Button("Generate Image",css=css_mt)
|
| 93 |
with gr.TabItem("Guided Diffusion"):
|
| 94 |
gr.Markdown("Guided Diffusion models produce superb quality results. V-Diffusion is its latest implementation")
|
| 95 |
steps_diff = gr.inputs.Slider(label="Steps - more steps can increase quality but will take longer to generate",default=40,maximum=80,minimum=1,step=1)
|
| 96 |
images_diff = gr.inputs.Slider(label="Number of images in parallel", default=2, maximum=4, minimum=1, step=1)
|
| 97 |
weight = gr.inputs.Slider(label="Weight - how closely the image should resemble the prompt", default=5, maximum=15, minimum=0, step=1)
|
| 98 |
clip = gr.inputs.Checkbox(label="CLIP Guided - improves coherence with complex prompts, makes it slower")
|
| 99 |
+
get_image_diffusion = gr.Button("Generate Image",css=css_mt)
|
| 100 |
with gr.Row():
|
| 101 |
with gr.Tabs():
|
| 102 |
#with gr.TabItem("Image output"):
|