Spaces:
Running
on
Zero
Running
on
Zero
Update inferencer.py
Browse files- inferencer.py +7 -0
inferencer.py
CHANGED
|
@@ -315,8 +315,15 @@ class UniPicV2Inferencer:
|
|
| 315 |
embeds = self._process_inputs(inputs, num_queries)
|
| 316 |
|
| 317 |
generator = torch.Generator(device=self.device).manual_seed(seed)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 318 |
|
| 319 |
edited_image = self.pipeline(
|
|
|
|
| 320 |
image=image,
|
| 321 |
prompt_embeds=embeds["prompt_embeds"].to(self.device),
|
| 322 |
pooled_prompt_embeds=embeds["pooled_prompt_embeds"].to(self.device),
|
|
|
|
| 315 |
embeds = self._process_inputs(inputs, num_queries)
|
| 316 |
|
| 317 |
generator = torch.Generator(device=self.device).manual_seed(seed)
|
| 318 |
+
|
| 319 |
+
latents = torch.randn(
|
| 320 |
+
1, self.pipeline.transformer.config.in_channels,
|
| 321 |
+
height // self.pipeline.vae_scale_factor,
|
| 322 |
+
width // self.pipeline.vae_scale_factor,
|
| 323 |
+
device=self.device, dtype=self.pipeline.transformer.dtype)
|
| 324 |
|
| 325 |
edited_image = self.pipeline(
|
| 326 |
+
latents=latents,
|
| 327 |
image=image,
|
| 328 |
prompt_embeds=embeds["prompt_embeds"].to(self.device),
|
| 329 |
pooled_prompt_embeds=embeds["pooled_prompt_embeds"].to(self.device),
|