Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -12,7 +12,7 @@ from previewer.modules import Previewer
|
|
| 12 |
|
| 13 |
from gallery_history import fetch_gallery_history, show_gallery_history
|
| 14 |
|
| 15 |
-
os.environ["TOKENIZERS_PARALLELISM"] =
|
| 16 |
|
| 17 |
DESCRIPTION = "# Würstchen"
|
| 18 |
DESCRIPTION += "\n<p style=\"text-align: center\"><a href='https://huggingface.co/warp-ai/wuerstchen' target='_blank'>Würstchen</a> is a new fast and efficient high resolution text-to-image architecture and model</p>"
|
|
@@ -29,12 +29,8 @@ PREVIEW_IMAGES = True
|
|
| 29 |
dtype = torch.float16
|
| 30 |
device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
|
| 31 |
if torch.cuda.is_available():
|
| 32 |
-
prior_pipeline = WuerstchenPriorPipeline.from_pretrained(
|
| 33 |
-
|
| 34 |
-
)
|
| 35 |
-
decoder_pipeline = WuerstchenDecoderPipeline.from_pretrained(
|
| 36 |
-
"warp-ai/wuerstchen", torch_dtype=dtype
|
| 37 |
-
)
|
| 38 |
if ENABLE_CPU_OFFLOAD:
|
| 39 |
prior_pipeline.enable_model_cpu_offload()
|
| 40 |
decoder_pipeline.enable_model_cpu_offload()
|
|
@@ -43,12 +39,8 @@ if torch.cuda.is_available():
|
|
| 43 |
decoder_pipeline.to(device)
|
| 44 |
|
| 45 |
if USE_TORCH_COMPILE:
|
| 46 |
-
prior_pipeline.prior = torch.compile(
|
| 47 |
-
|
| 48 |
-
)
|
| 49 |
-
decoder_pipeline.decoder = torch.compile(
|
| 50 |
-
decoder_pipeline.decoder, mode="reduce-overhead", fullgraph=True
|
| 51 |
-
)
|
| 52 |
|
| 53 |
if PREVIEW_IMAGES:
|
| 54 |
previewer = Previewer()
|
|
@@ -228,18 +220,18 @@ with gr.Blocks(css="style.css") as demo:
|
|
| 228 |
history = show_gallery_history()
|
| 229 |
|
| 230 |
inputs = [
|
| 231 |
-
|
| 232 |
-
|
| 233 |
-
|
| 234 |
-
|
| 235 |
-
|
| 236 |
-
|
| 237 |
-
|
| 238 |
-
|
| 239 |
-
|
| 240 |
-
|
| 241 |
-
|
| 242 |
-
|
| 243 |
]
|
| 244 |
prompt.submit(
|
| 245 |
fn=randomize_seed_fn,
|
|
|
|
| 12 |
|
| 13 |
from gallery_history import fetch_gallery_history, show_gallery_history
|
| 14 |
|
| 15 |
+
os.environ["TOKENIZERS_PARALLELISM"] = 'false'
|
| 16 |
|
| 17 |
DESCRIPTION = "# Würstchen"
|
| 18 |
DESCRIPTION += "\n<p style=\"text-align: center\"><a href='https://huggingface.co/warp-ai/wuerstchen' target='_blank'>Würstchen</a> is a new fast and efficient high resolution text-to-image architecture and model</p>"
|
|
|
|
| 29 |
dtype = torch.float16
|
| 30 |
device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
|
| 31 |
if torch.cuda.is_available():
|
| 32 |
+
prior_pipeline = WuerstchenPriorPipeline.from_pretrained("warp-ai/wuerstchen-prior", torch_dtype=dtype)
|
| 33 |
+
decoder_pipeline = WuerstchenDecoderPipeline.from_pretrained("warp-ai/wuerstchen", torch_dtype=dtype)
|
|
|
|
|
|
|
|
|
|
|
|
|
| 34 |
if ENABLE_CPU_OFFLOAD:
|
| 35 |
prior_pipeline.enable_model_cpu_offload()
|
| 36 |
decoder_pipeline.enable_model_cpu_offload()
|
|
|
|
| 39 |
decoder_pipeline.to(device)
|
| 40 |
|
| 41 |
if USE_TORCH_COMPILE:
|
| 42 |
+
prior_pipeline.prior = torch.compile(prior_pipeline.prior, mode="reduce-overhead", fullgraph=True)
|
| 43 |
+
decoder_pipeline.decoder = torch.compile(decoder_pipeline.decoder, mode="reduce-overhead", fullgraph=True)
|
|
|
|
|
|
|
|
|
|
|
|
|
| 44 |
|
| 45 |
if PREVIEW_IMAGES:
|
| 46 |
previewer = Previewer()
|
|
|
|
| 220 |
history = show_gallery_history()
|
| 221 |
|
| 222 |
inputs = [
|
| 223 |
+
prompt,
|
| 224 |
+
negative_prompt,
|
| 225 |
+
seed,
|
| 226 |
+
width,
|
| 227 |
+
height,
|
| 228 |
+
prior_num_inference_steps,
|
| 229 |
+
# prior_timesteps,
|
| 230 |
+
prior_guidance_scale,
|
| 231 |
+
decoder_num_inference_steps,
|
| 232 |
+
# decoder_timesteps,
|
| 233 |
+
decoder_guidance_scale,
|
| 234 |
+
num_images_per_prompt,
|
| 235 |
]
|
| 236 |
prompt.submit(
|
| 237 |
fn=randomize_seed_fn,
|