Spaces:
Paused
Paused
As input by default
Browse files
app.py
CHANGED
|
@@ -476,7 +476,7 @@ with gr.Blocks() as interface:
|
|
| 476 |
prompt_hint = gr.HTML("You can use a <a href='"'https://huggingface.co/spaces/MaziyarPanahi/llava-llama-3-8b'"'>LlaVa space</a> to auto-generate the description of your image.")
|
| 477 |
upscale = gr.Radio([["x1", 1], ["x2", 2], ["x3", 3], ["x4", 4], ["x5", 5], ["x6", 6], ["x7", 7], ["x8", 8]], label="Upscale factor", info="Resolution x1 to x8", value=2, interactive=True)
|
| 478 |
allocation = gr.Radio([["1 min", 1], ["2 min", 2], ["3 min", 3], ["4 min", 4], ["5 min", 5], ["6 min", 6], ["7 min", 7], ["8 min", 8], ["9 min (discouraged)", 9], ["10 min (discouraged)", 10]], label="GPU allocation time", info="lower=May abort run, higher=Quota penalty for next runs", value=6, interactive=True)
|
| 479 |
-
output_format = gr.Radio([["As input", "input"], ["*.png", "png"], ["*.webp", "webp"], ["*.jpeg", "jpeg"], ["*.gif", "gif"], ["*.bmp", "bmp"]], label="Image format for result", info="File extention", value="
|
| 480 |
|
| 481 |
with gr.Accordion("Pre-denoising (optional)", open=False):
|
| 482 |
gamma_correction = gr.Slider(label="Gamma Correction", info = "lower=lighter, higher=darker", minimum=0.1, maximum=2.0, value=1.0, step=0.1)
|
|
@@ -613,7 +613,7 @@ with gr.Blocks() as interface:
|
|
| 613 |
False,
|
| 614 |
0.,
|
| 615 |
"v0-Q",
|
| 616 |
-
"
|
| 617 |
5
|
| 618 |
],
|
| 619 |
[
|
|
@@ -644,7 +644,7 @@ with gr.Blocks() as interface:
|
|
| 644 |
False,
|
| 645 |
0.,
|
| 646 |
"v0-Q",
|
| 647 |
-
"
|
| 648 |
4
|
| 649 |
],
|
| 650 |
[
|
|
@@ -675,7 +675,7 @@ with gr.Blocks() as interface:
|
|
| 675 |
False,
|
| 676 |
0.,
|
| 677 |
"v0-Q",
|
| 678 |
-
"
|
| 679 |
4
|
| 680 |
],
|
| 681 |
[
|
|
@@ -706,7 +706,7 @@ with gr.Blocks() as interface:
|
|
| 706 |
False,
|
| 707 |
0.,
|
| 708 |
"v0-Q",
|
| 709 |
-
"
|
| 710 |
4
|
| 711 |
],
|
| 712 |
],
|
|
|
|
| 476 |
prompt_hint = gr.HTML("You can use a <a href='"'https://huggingface.co/spaces/MaziyarPanahi/llava-llama-3-8b'"'>LlaVa space</a> to auto-generate the description of your image.")
|
| 477 |
upscale = gr.Radio([["x1", 1], ["x2", 2], ["x3", 3], ["x4", 4], ["x5", 5], ["x6", 6], ["x7", 7], ["x8", 8]], label="Upscale factor", info="Resolution x1 to x8", value=2, interactive=True)
|
| 478 |
allocation = gr.Radio([["1 min", 1], ["2 min", 2], ["3 min", 3], ["4 min", 4], ["5 min", 5], ["6 min", 6], ["7 min", 7], ["8 min", 8], ["9 min (discouraged)", 9], ["10 min (discouraged)", 10]], label="GPU allocation time", info="lower=May abort run, higher=Quota penalty for next runs", value=6, interactive=True)
|
| 479 |
+
output_format = gr.Radio([["As input", "input"], ["*.png", "png"], ["*.webp", "webp"], ["*.jpeg", "jpeg"], ["*.gif", "gif"], ["*.bmp", "bmp"]], label="Image format for result", info="File extention", value="input", interactive=True)
|
| 480 |
|
| 481 |
with gr.Accordion("Pre-denoising (optional)", open=False):
|
| 482 |
gamma_correction = gr.Slider(label="Gamma Correction", info = "lower=lighter, higher=darker", minimum=0.1, maximum=2.0, value=1.0, step=0.1)
|
|
|
|
| 613 |
False,
|
| 614 |
0.,
|
| 615 |
"v0-Q",
|
| 616 |
+
"input",
|
| 617 |
5
|
| 618 |
],
|
| 619 |
[
|
|
|
|
| 644 |
False,
|
| 645 |
0.,
|
| 646 |
"v0-Q",
|
| 647 |
+
"input",
|
| 648 |
4
|
| 649 |
],
|
| 650 |
[
|
|
|
|
| 675 |
False,
|
| 676 |
0.,
|
| 677 |
"v0-Q",
|
| 678 |
+
"input",
|
| 679 |
4
|
| 680 |
],
|
| 681 |
[
|
|
|
|
| 706 |
False,
|
| 707 |
0.,
|
| 708 |
"v0-Q",
|
| 709 |
+
"input",
|
| 710 |
4
|
| 711 |
],
|
| 712 |
],
|