Update app.py
Browse files
app.py
CHANGED
|
@@ -10,6 +10,7 @@ if not torch.cuda.is_available():
|
|
| 10 |
DESCRIPTION += "\n<p>你现在运行在CPU上 但是此项目只支持GPU.</p>"
|
| 11 |
|
| 12 |
MAX_SEED = np.iinfo(np.int32).max
|
|
|
|
| 13 |
MAX_IMAGE_SIZE = 4096
|
| 14 |
|
| 15 |
if torch.cuda.is_available():
|
|
@@ -63,8 +64,8 @@ def infer(
|
|
| 63 |
return image, seed
|
| 64 |
|
| 65 |
examples = [
|
| 66 |
-
"
|
| 67 |
-
"
|
| 68 |
]
|
| 69 |
|
| 70 |
css = '''
|
|
@@ -139,6 +140,14 @@ with gr.Blocks(css=css) as demo:
|
|
| 139 |
value=28,
|
| 140 |
)
|
| 141 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 142 |
use_negative_prompt.change(
|
| 143 |
fn=lambda x: gr.update(visible=x),
|
| 144 |
inputs=use_negative_prompt,
|
|
|
|
| 10 |
DESCRIPTION += "\n<p>你现在运行在CPU上 但是此项目只支持GPU.</p>"
|
| 11 |
|
| 12 |
MAX_SEED = np.iinfo(np.int32).max
|
| 13 |
+
CACHE_EXAMPLES = torch.cuda.is_available() and os.getenv("CACHE_EXAMPLES", "1") == "1"
|
| 14 |
MAX_IMAGE_SIZE = 4096
|
| 15 |
|
| 16 |
if torch.cuda.is_available():
|
|
|
|
| 64 |
return image, seed
|
| 65 |
|
| 66 |
examples = [
|
| 67 |
+
"nahida (genshin impact)",
|
| 68 |
+
"klee (genshin impact)",
|
| 69 |
]
|
| 70 |
|
| 71 |
css = '''
|
|
|
|
| 140 |
value=28,
|
| 141 |
)
|
| 142 |
|
| 143 |
+
gr.Examples(
|
| 144 |
+
examples=examples,
|
| 145 |
+
inputs=prompt,
|
| 146 |
+
outputs=[result, seed],
|
| 147 |
+
fn=generate,
|
| 148 |
+
cache_examples=CACHE_EXAMPLES,
|
| 149 |
+
)
|
| 150 |
+
|
| 151 |
use_negative_prompt.change(
|
| 152 |
fn=lambda x: gr.update(visible=x),
|
| 153 |
inputs=use_negative_prompt,
|