| import math |
| import tempfile |
| from typing import List, Optional |
|
|
| import numpy as np |
| import PIL.Image |
| import torch |
| from accelerate import Accelerator |
| from torchvision import transforms |
| from tqdm.auto import tqdm |
| from transformers import CLIPTextModel, CLIPTokenizer |
|
|
| from diffusers import AutoencoderKL, DiffusionPipeline, DPMSolverMultistepScheduler, UNet2DConditionModel |
| from diffusers.loaders import AttnProcsLayers, LoraLoaderMixin |
| from diffusers.models.attention_processor import ( |
| AttnAddedKVProcessor, |
| AttnAddedKVProcessor2_0, |
| LoRAAttnAddedKVProcessor, |
| LoRAAttnProcessor, |
| LoRAAttnProcessor2_0, |
| SlicedAttnAddedKVProcessor, |
| ) |
| from diffusers.optimization import get_scheduler |
|
|
|
|
| class SdeDragPipeline(DiffusionPipeline): |
| r""" |
| Pipeline for image drag-and-drop editing using stochastic differential equations: https://arxiv.org/abs/2311.01410. |
| Please refer to the [official repository](https://github.com/ML-GSAI/SDE-Drag) for more information. |
| |
| This model inherits from [`DiffusionPipeline`]. Check the superclass documentation for the generic methods the |
| library implements for all the pipelines (such as downloading or saving, running on a particular device, etc.) |
| |
| Args: |
| vae ([`AutoencoderKL`]): |
| Variational Auto-Encoder (VAE) Model to encode and decode images to and from latent representations. |
| text_encoder ([`CLIPTextModel`]): |
| Frozen text-encoder. Stable Diffusion uses the text portion of |
| [CLIP](https://huggingface.co/docs/transformers/model_doc/clip#transformers.CLIPTextModel), specifically |
| the [clip-vit-large-patch14](https://huggingface.co/openai/clip-vit-large-patch14) variant. |
| tokenizer (`CLIPTokenizer`): |
| Tokenizer of class |
| [CLIPTokenizer](https://huggingface.co/docs/transformers/v4.21.0/en/model_doc/clip#transformers.CLIPTokenizer). |
| unet ([`UNet2DConditionModel`]): Conditional U-Net architecture to denoise the encoded image latents. |
| scheduler ([`SchedulerMixin`]): |
| A scheduler to be used in combination with `unet` to denoise the encoded image latents. Please use |
| [`DDIMScheduler`]. |
| """ |
|
|
| def __init__( |
| self, |
| vae: AutoencoderKL, |
| text_encoder: CLIPTextModel, |
| tokenizer: CLIPTokenizer, |
| unet: UNet2DConditionModel, |
| scheduler: DPMSolverMultistepScheduler, |
| ): |
| super().__init__() |
|
|
| self.register_modules(vae=vae, text_encoder=text_encoder, tokenizer=tokenizer, unet=unet, scheduler=scheduler) |
|
|
| @torch.no_grad() |
| def __call__( |
| self, |
| prompt: str, |
| image: PIL.Image.Image, |
| mask_image: PIL.Image.Image, |
| source_points: List[List[int]], |
| target_points: List[List[int]], |
| t0: Optional[float] = 0.6, |
| steps: Optional[int] = 200, |
| step_size: Optional[int] = 2, |
| image_scale: Optional[float] = 0.3, |
| adapt_radius: Optional[int] = 5, |
| min_lora_scale: Optional[float] = 0.5, |
| generator: Optional[torch.Generator] = None, |
| ): |
| r""" |
| Function invoked when calling the pipeline for image editing. |
| Args: |
| prompt (`str`, *required*): |
| The prompt to guide the image editing. |
| image (`PIL.Image.Image`, *required*): |
| Which will be edited, parts of the image will be masked out with `mask_image` and edited |
| according to `prompt`. |
| mask_image (`PIL.Image.Image`, *required*): |
| To mask `image`. White pixels in the mask will be edited, while black pixels will be preserved. |
| source_points (`List[List[int]]`, *required*): |
| Used to mark the starting positions of drag editing in the image, with each pixel represented as a |
| `List[int]` of length 2. |
| target_points (`List[List[int]]`, *required*): |
| Used to mark the target positions of drag editing in the image, with each pixel represented as a |
| `List[int]` of length 2. |
| t0 (`float`, *optional*, defaults to 0.6): |
| The time parameter. Higher t0 improves the fidelity while lowering the faithfulness of the edited images |
| and vice versa. |
| steps (`int`, *optional*, defaults to 200): |
| The number of sampling iterations. |
| step_size (`int`, *optional*, defaults to 2): |
| The drag diatance of each drag step. |
| image_scale (`float`, *optional*, defaults to 0.3): |
| To avoid duplicating the content, use image_scale to perturbs the source. |
| adapt_radius (`int`, *optional*, defaults to 5): |
| The size of the region for copy and paste operations during each step of the drag process. |
| min_lora_scale (`float`, *optional*, defaults to 0.5): |
| A lora scale that will be applied to all LoRA layers of the text encoder if LoRA layers are loaded. |
| min_lora_scale specifies the minimum LoRA scale during the image drag-editing process. |
| generator ('torch.Generator', *optional*, defaults to None): |
| To make generation deterministic(https://pytorch.org/docs/stable/generated/torch.Generator.html). |
| Examples: |
| ```py |
| >>> import PIL |
| >>> import torch |
| >>> from diffusers import DDIMScheduler, DiffusionPipeline |
| |
| >>> # Load the pipeline |
| >>> model_path = "runwayml/stable-diffusion-v1-5" |
| >>> scheduler = DDIMScheduler.from_pretrained(model_path, subfolder="scheduler") |
| >>> pipe = DiffusionPipeline.from_pretrained(model_path, scheduler=scheduler, custom_pipeline="sde_drag") |
| >>> pipe.to('cuda') |
| |
| >>> # To save GPU memory, torch.float16 can be used, but it may compromise image quality. |
| >>> # If not training LoRA, please avoid using torch.float16 |
| >>> # pipe.to(torch.float16) |
| |
| >>> # Provide prompt, image, mask image, and the starting and target points for drag editing. |
| >>> prompt = "prompt of the image" |
| >>> image = PIL.Image.open('/path/to/image') |
| >>> mask_image = PIL.Image.open('/path/to/mask_image') |
| >>> source_points = [[123, 456]] |
| >>> target_points = [[234, 567]] |
| |
| >>> # train_lora is optional, and in most cases, using train_lora can better preserve consistency with the original image. |
| >>> pipe.train_lora(prompt, image) |
| |
| >>> output = pipe(prompt, image, mask_image, source_points, target_points) |
| >>> output_image = PIL.Image.fromarray(output) |
| >>> output_image.save("./output.png") |
| ``` |
| """ |
|
|
| self.scheduler.set_timesteps(steps) |
|
|
| noise_scale = (1 - image_scale**2) ** (0.5) |
|
|
| text_embeddings = self._get_text_embed(prompt) |
| uncond_embeddings = self._get_text_embed([""]) |
| text_embeddings = torch.cat([uncond_embeddings, text_embeddings]) |
|
|
| latent = self._get_img_latent(image) |
|
|
| mask = mask_image.resize((latent.shape[3], latent.shape[2])) |
| mask = torch.tensor(np.array(mask)) |
| mask = mask.unsqueeze(0).expand_as(latent).to(self.device) |
|
|
| source_points = torch.tensor(source_points).div(torch.tensor([8]), rounding_mode="trunc") |
| target_points = torch.tensor(target_points).div(torch.tensor([8]), rounding_mode="trunc") |
|
|
| distance = target_points - source_points |
| distance_norm_max = torch.norm(distance.float(), dim=1, keepdim=True).max() |
|
|
| if distance_norm_max <= step_size: |
| drag_num = 1 |
| else: |
| drag_num = distance_norm_max.div(torch.tensor([step_size]), rounding_mode="trunc") |
| if (distance_norm_max / drag_num - step_size).abs() > ( |
| distance_norm_max / (drag_num + 1) - step_size |
| ).abs(): |
| drag_num += 1 |
|
|
| latents = [] |
| for i in tqdm(range(int(drag_num)), desc="SDE Drag"): |
| source_new = source_points + (i / drag_num * distance).to(torch.int) |
| target_new = source_points + ((i + 1) / drag_num * distance).to(torch.int) |
|
|
| latent, noises, hook_latents, lora_scales, cfg_scales = self._forward( |
| latent, steps, t0, min_lora_scale, text_embeddings, generator |
| ) |
| latent = self._copy_and_paste( |
| latent, |
| source_new, |
| target_new, |
| adapt_radius, |
| latent.shape[2] - 1, |
| latent.shape[3] - 1, |
| image_scale, |
| noise_scale, |
| generator, |
| ) |
| latent = self._backward( |
| latent, mask, steps, t0, noises, hook_latents, lora_scales, cfg_scales, text_embeddings, generator |
| ) |
|
|
| latents.append(latent) |
|
|
| result_image = 1 / 0.18215 * latents[-1] |
|
|
| with torch.no_grad(): |
| result_image = self.vae.decode(result_image).sample |
|
|
| result_image = (result_image / 2 + 0.5).clamp(0, 1) |
| result_image = result_image.cpu().permute(0, 2, 3, 1).numpy()[0] |
| result_image = (result_image * 255).astype(np.uint8) |
|
|
| return result_image |
|
|
| def train_lora(self, prompt, image, lora_step=100, lora_rank=16, generator=None): |
| accelerator = Accelerator(gradient_accumulation_steps=1, mixed_precision="fp16") |
|
|
| self.vae.requires_grad_(False) |
| self.text_encoder.requires_grad_(False) |
| self.unet.requires_grad_(False) |
|
|
| unet_lora_attn_procs = {} |
| for name, attn_processor in self.unet.attn_processors.items(): |
| cross_attention_dim = None if name.endswith("attn1.processor") else self.unet.config.cross_attention_dim |
| if name.startswith("mid_block"): |
| hidden_size = self.unet.config.block_out_channels[-1] |
| elif name.startswith("up_blocks"): |
| block_id = int(name[len("up_blocks.")]) |
| hidden_size = list(reversed(self.unet.config.block_out_channels))[block_id] |
| elif name.startswith("down_blocks"): |
| block_id = int(name[len("down_blocks.")]) |
| hidden_size = self.unet.config.block_out_channels[block_id] |
| else: |
| raise NotImplementedError("name must start with up_blocks, mid_blocks, or down_blocks") |
|
|
| if isinstance(attn_processor, (AttnAddedKVProcessor, SlicedAttnAddedKVProcessor, AttnAddedKVProcessor2_0)): |
| lora_attn_processor_class = LoRAAttnAddedKVProcessor |
| else: |
| lora_attn_processor_class = ( |
| LoRAAttnProcessor2_0 |
| if hasattr(torch.nn.functional, "scaled_dot_product_attention") |
| else LoRAAttnProcessor |
| ) |
| unet_lora_attn_procs[name] = lora_attn_processor_class( |
| hidden_size=hidden_size, cross_attention_dim=cross_attention_dim, rank=lora_rank |
| ) |
|
|
| self.unet.set_attn_processor(unet_lora_attn_procs) |
| unet_lora_layers = AttnProcsLayers(self.unet.attn_processors) |
| params_to_optimize = unet_lora_layers.parameters() |
|
|
| optimizer = torch.optim.AdamW( |
| params_to_optimize, |
| lr=2e-4, |
| betas=(0.9, 0.999), |
| weight_decay=1e-2, |
| eps=1e-08, |
| ) |
|
|
| lr_scheduler = get_scheduler( |
| "constant", |
| optimizer=optimizer, |
| num_warmup_steps=0, |
| num_training_steps=lora_step, |
| num_cycles=1, |
| power=1.0, |
| ) |
|
|
| unet_lora_layers = accelerator.prepare_model(unet_lora_layers) |
| optimizer = accelerator.prepare_optimizer(optimizer) |
| lr_scheduler = accelerator.prepare_scheduler(lr_scheduler) |
|
|
| with torch.no_grad(): |
| text_inputs = self._tokenize_prompt(prompt, tokenizer_max_length=None) |
| text_embedding = self._encode_prompt( |
| text_inputs.input_ids, text_inputs.attention_mask, text_encoder_use_attention_mask=False |
| ) |
|
|
| image_transforms = transforms.Compose( |
| [ |
| transforms.ToTensor(), |
| transforms.Normalize([0.5], [0.5]), |
| ] |
| ) |
|
|
| image = image_transforms(image).to(self.device, dtype=self.vae.dtype) |
| image = image.unsqueeze(dim=0) |
| latents_dist = self.vae.encode(image).latent_dist |
|
|
| for _ in tqdm(range(lora_step), desc="Train LoRA"): |
| self.unet.train() |
| model_input = latents_dist.sample() * self.vae.config.scaling_factor |
|
|
| |
| noise = torch.randn( |
| model_input.size(), |
| dtype=model_input.dtype, |
| layout=model_input.layout, |
| device=model_input.device, |
| generator=generator, |
| ) |
| bsz, channels, height, width = model_input.shape |
|
|
| |
| timesteps = torch.randint( |
| 0, self.scheduler.config.num_train_timesteps, (bsz,), device=model_input.device, generator=generator |
| ) |
| timesteps = timesteps.long() |
|
|
| |
| |
| noisy_model_input = self.scheduler.add_noise(model_input, noise, timesteps) |
|
|
| |
| model_pred = self.unet(noisy_model_input, timesteps, text_embedding).sample |
|
|
| |
| if self.scheduler.config.prediction_type == "epsilon": |
| target = noise |
| elif self.scheduler.config.prediction_type == "v_prediction": |
| target = self.scheduler.get_velocity(model_input, noise, timesteps) |
| else: |
| raise ValueError(f"Unknown prediction type {self.scheduler.config.prediction_type}") |
|
|
| loss = torch.nn.functional.mse_loss(model_pred.float(), target.float(), reduction="mean") |
| accelerator.backward(loss) |
| optimizer.step() |
| lr_scheduler.step() |
| optimizer.zero_grad() |
|
|
| with tempfile.TemporaryDirectory() as save_lora_dir: |
| LoraLoaderMixin.save_lora_weights( |
| save_directory=save_lora_dir, |
| unet_lora_layers=unet_lora_layers, |
| text_encoder_lora_layers=None, |
| ) |
|
|
| self.unet.load_attn_procs(save_lora_dir) |
|
|
| def _tokenize_prompt(self, prompt, tokenizer_max_length=None): |
| if tokenizer_max_length is not None: |
| max_length = tokenizer_max_length |
| else: |
| max_length = self.tokenizer.model_max_length |
|
|
| text_inputs = self.tokenizer( |
| prompt, |
| truncation=True, |
| padding="max_length", |
| max_length=max_length, |
| return_tensors="pt", |
| ) |
|
|
| return text_inputs |
|
|
| def _encode_prompt(self, input_ids, attention_mask, text_encoder_use_attention_mask=False): |
| text_input_ids = input_ids.to(self.device) |
|
|
| if text_encoder_use_attention_mask: |
| attention_mask = attention_mask.to(self.device) |
| else: |
| attention_mask = None |
|
|
| prompt_embeds = self.text_encoder( |
| text_input_ids, |
| attention_mask=attention_mask, |
| ) |
| prompt_embeds = prompt_embeds[0] |
|
|
| return prompt_embeds |
|
|
| @torch.no_grad() |
| def _get_text_embed(self, prompt): |
| text_input = self.tokenizer( |
| prompt, |
| padding="max_length", |
| max_length=self.tokenizer.model_max_length, |
| truncation=True, |
| return_tensors="pt", |
| ) |
| text_embeddings = self.text_encoder(text_input.input_ids.to(self.device))[0] |
| return text_embeddings |
|
|
| def _copy_and_paste( |
| self, latent, source_new, target_new, adapt_radius, max_height, max_width, image_scale, noise_scale, generator |
| ): |
| def adaption_r(source, target, adapt_radius, max_height, max_width): |
| r_x_lower = min(adapt_radius, source[0], target[0]) |
| r_x_upper = min(adapt_radius, max_width - source[0], max_width - target[0]) |
| r_y_lower = min(adapt_radius, source[1], target[1]) |
| r_y_upper = min(adapt_radius, max_height - source[1], max_height - target[1]) |
| return r_x_lower, r_x_upper, r_y_lower, r_y_upper |
|
|
| for source_, target_ in zip(source_new, target_new): |
| r_x_lower, r_x_upper, r_y_lower, r_y_upper = adaption_r( |
| source_, target_, adapt_radius, max_height, max_width |
| ) |
|
|
| source_feature = latent[ |
| :, :, source_[1] - r_y_lower : source_[1] + r_y_upper, source_[0] - r_x_lower : source_[0] + r_x_upper |
| ].clone() |
|
|
| latent[ |
| :, :, source_[1] - r_y_lower : source_[1] + r_y_upper, source_[0] - r_x_lower : source_[0] + r_x_upper |
| ] = image_scale * source_feature + noise_scale * torch.randn( |
| latent.shape[0], |
| 4, |
| r_y_lower + r_y_upper, |
| r_x_lower + r_x_upper, |
| device=self.device, |
| generator=generator, |
| ) |
|
|
| latent[ |
| :, :, target_[1] - r_y_lower : target_[1] + r_y_upper, target_[0] - r_x_lower : target_[0] + r_x_upper |
| ] = source_feature * 1.1 |
| return latent |
|
|
| @torch.no_grad() |
| def _get_img_latent(self, image, height=None, weight=None): |
| data = image.convert("RGB") |
| if height is not None: |
| data = data.resize((weight, height)) |
| transform = transforms.ToTensor() |
| data = transform(data).unsqueeze(0) |
| data = (data * 2.0) - 1.0 |
| data = data.to(self.device, dtype=self.vae.dtype) |
| latent = self.vae.encode(data).latent_dist.sample() |
| latent = 0.18215 * latent |
| return latent |
|
|
| @torch.no_grad() |
| def _get_eps(self, latent, timestep, guidance_scale, text_embeddings, lora_scale=None): |
| latent_model_input = torch.cat([latent] * 2) if guidance_scale > 1.0 else latent |
| text_embeddings = text_embeddings if guidance_scale > 1.0 else text_embeddings.chunk(2)[1] |
|
|
| cross_attention_kwargs = None if lora_scale is None else {"scale": lora_scale} |
|
|
| with torch.no_grad(): |
| noise_pred = self.unet( |
| latent_model_input, |
| timestep, |
| encoder_hidden_states=text_embeddings, |
| cross_attention_kwargs=cross_attention_kwargs, |
| ).sample |
|
|
| if guidance_scale > 1.0: |
| noise_pred_uncond, noise_pred_text = noise_pred.chunk(2) |
| elif guidance_scale == 1.0: |
| noise_pred_text = noise_pred |
| noise_pred_uncond = 0.0 |
| else: |
| raise NotImplementedError(guidance_scale) |
| noise_pred = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) |
|
|
| return noise_pred |
|
|
| def _forward_sde( |
| self, timestep, sample, guidance_scale, text_embeddings, steps, eta=1.0, lora_scale=None, generator=None |
| ): |
| num_train_timesteps = len(self.scheduler) |
| alphas_cumprod = self.scheduler.alphas_cumprod |
| initial_alpha_cumprod = torch.tensor(1.0) |
|
|
| prev_timestep = timestep + num_train_timesteps // steps |
|
|
| alpha_prod_t = alphas_cumprod[timestep] if timestep >= 0 else initial_alpha_cumprod |
| alpha_prod_t_prev = alphas_cumprod[prev_timestep] |
|
|
| beta_prod_t_prev = 1 - alpha_prod_t_prev |
|
|
| x_prev = (alpha_prod_t_prev / alpha_prod_t) ** (0.5) * sample + (1 - alpha_prod_t_prev / alpha_prod_t) ** ( |
| 0.5 |
| ) * torch.randn( |
| sample.size(), dtype=sample.dtype, layout=sample.layout, device=self.device, generator=generator |
| ) |
| eps = self._get_eps(x_prev, prev_timestep, guidance_scale, text_embeddings, lora_scale) |
|
|
| sigma_t_prev = ( |
| eta |
| * (1 - alpha_prod_t) ** (0.5) |
| * (1 - alpha_prod_t_prev / (1 - alpha_prod_t_prev) * (1 - alpha_prod_t) / alpha_prod_t) ** (0.5) |
| ) |
|
|
| pred_original_sample = (x_prev - beta_prod_t_prev ** (0.5) * eps) / alpha_prod_t_prev ** (0.5) |
| pred_sample_direction_coeff = (1 - alpha_prod_t - sigma_t_prev**2) ** (0.5) |
|
|
| noise = ( |
| sample - alpha_prod_t ** (0.5) * pred_original_sample - pred_sample_direction_coeff * eps |
| ) / sigma_t_prev |
|
|
| return x_prev, noise |
|
|
| def _sample( |
| self, |
| timestep, |
| sample, |
| guidance_scale, |
| text_embeddings, |
| steps, |
| sde=False, |
| noise=None, |
| eta=1.0, |
| lora_scale=None, |
| generator=None, |
| ): |
| num_train_timesteps = len(self.scheduler) |
| alphas_cumprod = self.scheduler.alphas_cumprod |
| final_alpha_cumprod = torch.tensor(1.0) |
|
|
| eps = self._get_eps(sample, timestep, guidance_scale, text_embeddings, lora_scale) |
|
|
| prev_timestep = timestep - num_train_timesteps // steps |
|
|
| alpha_prod_t = alphas_cumprod[timestep] |
| alpha_prod_t_prev = alphas_cumprod[prev_timestep] if prev_timestep >= 0 else final_alpha_cumprod |
|
|
| beta_prod_t = 1 - alpha_prod_t |
|
|
| sigma_t = ( |
| eta |
| * ((1 - alpha_prod_t_prev) / (1 - alpha_prod_t)) ** (0.5) |
| * (1 - alpha_prod_t / alpha_prod_t_prev) ** (0.5) |
| if sde |
| else 0 |
| ) |
|
|
| pred_original_sample = (sample - beta_prod_t ** (0.5) * eps) / alpha_prod_t ** (0.5) |
| pred_sample_direction_coeff = (1 - alpha_prod_t_prev - sigma_t**2) ** (0.5) |
|
|
| noise = ( |
| torch.randn( |
| sample.size(), dtype=sample.dtype, layout=sample.layout, device=self.device, generator=generator |
| ) |
| if noise is None |
| else noise |
| ) |
| latent = ( |
| alpha_prod_t_prev ** (0.5) * pred_original_sample + pred_sample_direction_coeff * eps + sigma_t * noise |
| ) |
|
|
| return latent |
|
|
| def _forward(self, latent, steps, t0, lora_scale_min, text_embeddings, generator): |
| def scale_schedule(begin, end, n, length, type="linear"): |
| if type == "constant": |
| return end |
| elif type == "linear": |
| return begin + (end - begin) * n / length |
| elif type == "cos": |
| factor = (1 - math.cos(n * math.pi / length)) / 2 |
| return (1 - factor) * begin + factor * end |
| else: |
| raise NotImplementedError(type) |
|
|
| noises = [] |
| latents = [] |
| lora_scales = [] |
| cfg_scales = [] |
| latents.append(latent) |
| t0 = int(t0 * steps) |
| t_begin = steps - t0 |
|
|
| length = len(self.scheduler.timesteps[t_begin - 1 : -1]) - 1 |
| index = 1 |
| for t in self.scheduler.timesteps[t_begin:].flip(dims=[0]): |
| lora_scale = scale_schedule(1, lora_scale_min, index, length, type="cos") |
| cfg_scale = scale_schedule(1, 3.0, index, length, type="linear") |
| latent, noise = self._forward_sde( |
| t, latent, cfg_scale, text_embeddings, steps, lora_scale=lora_scale, generator=generator |
| ) |
|
|
| noises.append(noise) |
| latents.append(latent) |
| lora_scales.append(lora_scale) |
| cfg_scales.append(cfg_scale) |
| index += 1 |
| return latent, noises, latents, lora_scales, cfg_scales |
|
|
| def _backward( |
| self, latent, mask, steps, t0, noises, hook_latents, lora_scales, cfg_scales, text_embeddings, generator |
| ): |
| t0 = int(t0 * steps) |
| t_begin = steps - t0 |
|
|
| hook_latent = hook_latents.pop() |
| latent = torch.where(mask > 128, latent, hook_latent) |
| for t in self.scheduler.timesteps[t_begin - 1 : -1]: |
| latent = self._sample( |
| t, |
| latent, |
| cfg_scales.pop(), |
| text_embeddings, |
| steps, |
| sde=True, |
| noise=noises.pop(), |
| lora_scale=lora_scales.pop(), |
| generator=generator, |
| ) |
| hook_latent = hook_latents.pop() |
| latent = torch.where(mask > 128, latent, hook_latent) |
| return latent |
|
|