Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -60,12 +60,17 @@ def fill_image(prompt, image, model_selection, paste_back):
|
|
| 60 |
|
| 61 |
pipe = get_pipeline(model_selection)
|
| 62 |
|
|
|
|
| 63 |
(
|
| 64 |
prompt_embeds,
|
| 65 |
negative_prompt_embeds,
|
| 66 |
pooled_prompt_embeds,
|
| 67 |
negative_pooled_prompt_embeds,
|
| 68 |
) = pipe.encode_prompt(prompt, "cuda", True)
|
|
|
|
|
|
|
|
|
|
|
|
|
| 69 |
|
| 70 |
source = image["background"]
|
| 71 |
mask = image["layers"][0]
|
|
@@ -251,12 +256,17 @@ def infer(image, width, height, overlap_percentage, num_inference_steps, resize_
|
|
| 251 |
final_prompt = f"{prompt_input}, high quality, 4k"
|
| 252 |
|
| 253 |
pipe = get_pipeline(model_selection)
|
|
|
|
| 254 |
(
|
| 255 |
prompt_embeds,
|
| 256 |
negative_prompt_embeds,
|
| 257 |
pooled_prompt_embeds,
|
| 258 |
negative_pooled_prompt_embeds,
|
| 259 |
) = pipe.encode_prompt(final_prompt, "cuda", True)
|
|
|
|
|
|
|
|
|
|
|
|
|
| 260 |
|
| 261 |
for output_image in pipe(
|
| 262 |
prompt_embeds=prompt_embeds,
|
|
|
|
| 60 |
|
| 61 |
pipe = get_pipeline(model_selection)
|
| 62 |
|
| 63 |
+
# Encode the prompt and ensure the embeddings are in float16
|
| 64 |
(
|
| 65 |
prompt_embeds,
|
| 66 |
negative_prompt_embeds,
|
| 67 |
pooled_prompt_embeds,
|
| 68 |
negative_pooled_prompt_embeds,
|
| 69 |
) = pipe.encode_prompt(prompt, "cuda", True)
|
| 70 |
+
prompt_embeds = prompt_embeds.to(torch.float16)
|
| 71 |
+
negative_prompt_embeds = negative_prompt_embeds.to(torch.float16)
|
| 72 |
+
pooled_prompt_embeds = pooled_prompt_embeds.to(torch.float16)
|
| 73 |
+
negative_pooled_prompt_embeds = negative_pooled_prompt_embeds.to(torch.float16)
|
| 74 |
|
| 75 |
source = image["background"]
|
| 76 |
mask = image["layers"][0]
|
|
|
|
| 256 |
final_prompt = f"{prompt_input}, high quality, 4k"
|
| 257 |
|
| 258 |
pipe = get_pipeline(model_selection)
|
| 259 |
+
# Encode the prompt and ensure the embeddings are in float16
|
| 260 |
(
|
| 261 |
prompt_embeds,
|
| 262 |
negative_prompt_embeds,
|
| 263 |
pooled_prompt_embeds,
|
| 264 |
negative_pooled_prompt_embeds,
|
| 265 |
) = pipe.encode_prompt(final_prompt, "cuda", True)
|
| 266 |
+
prompt_embeds = prompt_embeds.to(torch.float16)
|
| 267 |
+
negative_prompt_embeds = negative_prompt_embeds.to(torch.float16)
|
| 268 |
+
pooled_prompt_embeds = pooled_prompt_embeds.to(torch.float16)
|
| 269 |
+
negative_pooled_prompt_embeds = negative_pooled_prompt_embeds.to(torch.float16)
|
| 270 |
|
| 271 |
for output_image in pipe(
|
| 272 |
prompt_embeds=prompt_embeds,
|