| |
|
|
| import imp |
| import numpy as np |
| import torch |
| import random |
| from PIL import Image, ImageDraw, ImageFont |
| import copy |
| from typing import Optional, Union, Tuple, List, Callable, Dict, Any |
| from tqdm.notebook import tqdm |
| from diffusers.utils import BaseOutput, logging |
| from diffusers.models.embeddings import TimestepEmbedding, Timesteps |
| from diffusers.models.unet_2d_blocks import ( |
| CrossAttnDownBlock2D, |
| CrossAttnUpBlock2D, |
| DownBlock2D, |
| UNetMidBlock2DCrossAttn, |
| UpBlock2D, |
| get_down_block, |
| get_up_block, |
| ) |
| from diffusers.models.unet_2d_condition import UNet2DConditionOutput |
| from copy import deepcopy |
| import json |
|
|
| import inspect |
| import os |
| import warnings |
| from typing import Any, Callable, Dict, List, Optional, Tuple, Union |
|
|
| import numpy as np |
| import PIL.Image |
| import torch |
| import torch.nn.functional as F |
| from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer |
|
|
| from diffusers.image_processor import VaeImageProcessor |
| from diffusers.loaders import LoraLoaderMixin, TextualInversionLoaderMixin |
| from diffusers.models import AutoencoderKL, ControlNetModel, UNet2DConditionModel |
| from diffusers.schedulers import KarrasDiffusionSchedulers |
| from diffusers.utils.torch_utils import is_compiled_module |
|
|
| from diffusers.pipelines.stable_diffusion import StableDiffusionPipelineOutput |
| from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker |
| from diffusers.pipelines.controlnet.multicontrolnet import MultiControlNetModel |
| from tqdm import tqdm |
| import time |
|
|
| from dataclasses import dataclass |
| from typing import Any, Dict, List, Optional, Tuple, Union |
|
|
| import torch |
| import torch.nn as nn |
| import torch.utils.checkpoint |
|
|
| from diffusers.models.embeddings import ( |
| GaussianFourierProjection, |
| ImageHintTimeEmbedding, |
| ImageProjection, |
| ImageTimeEmbedding, |
| TextImageProjection, |
| TextImageTimeEmbedding, |
| TextTimeEmbedding, |
| TimestepEmbedding, |
| Timesteps, |
| ) |
| from diffusers.models.modeling_utils import ModelMixin |
| from diffusers.models.unet_2d_blocks import ( |
| UNetMidBlock2D, |
| UNetMidBlock2DCrossAttn, |
| UNetMidBlock2DSimpleCrossAttn, |
| get_down_block, |
| get_up_block, |
| ) |
| |
| from diffusers.image_processor import PipelineImageInput, VaeImageProcessor |
| from diffusers.loaders import ( |
| FromSingleFileMixin, |
| IPAdapterMixin, |
| StableDiffusionXLLoraLoaderMixin, |
| TextualInversionLoaderMixin, |
| ) |
| from diffusers.models import AutoencoderKL, ImageProjection, UNet2DConditionModel |
| from diffusers.models.attention_processor import ( |
| AttnProcessor2_0, |
| FusedAttnProcessor2_0, |
| XFormersAttnProcessor, |
| ) |
| from diffusers.models.lora import adjust_lora_scale_text_encoder |
| from diffusers.schedulers import KarrasDiffusionSchedulers |
| from diffusers.utils import ( |
| USE_PEFT_BACKEND, |
| deprecate, |
| is_invisible_watermark_available, |
| is_torch_xla_available, |
| logging, |
| replace_example_docstring, |
| scale_lora_layers, |
| unscale_lora_layers, |
| ) |
| from diffusers.utils.torch_utils import randn_tensor |
| |
| from diffusers.pipelines.stable_diffusion_xl.pipeline_output import StableDiffusionXLPipelineOutput |
|
|
| from typing import Any, Dict, List |
|
|
| from diffusers.configuration_utils import ConfigMixin, register_to_config |
| from diffusers.utils import CONFIG_NAME |
|
|
|
|
| class PipelineCallback(ConfigMixin): |
|
|
| config_name = CONFIG_NAME |
|
|
| @register_to_config |
| def __init__(self, cutoff_step_ratio=1.0, cutoff_step_index=None): |
| super().__init__() |
|
|
| if (cutoff_step_ratio is None and cutoff_step_index is None) or ( |
| cutoff_step_ratio is not None and cutoff_step_index is not None |
| ): |
| raise ValueError("Either cutoff_step_ratio or cutoff_step_index should be provided, not both or none.") |
|
|
| if cutoff_step_ratio is not None and ( |
| not isinstance(cutoff_step_ratio, float) or not (0.0 <= cutoff_step_ratio <= 1.0) |
| ): |
| raise ValueError("cutoff_step_ratio must be a float between 0.0 and 1.0.") |
|
|
| @property |
| def tensor_inputs(self) -> List[str]: |
| raise NotImplementedError(f"You need to set the attribute `tensor_inputs` for {self.__class__}") |
|
|
| def callback_fn(self, pipeline, step_index, timesteps, callback_kwargs) -> Dict[str, Any]: |
| raise NotImplementedError(f"You need to implement the method `callback_fn` for {self.__class__}") |
|
|
| def __call__(self, pipeline, step_index, timestep, callback_kwargs) -> Dict[str, Any]: |
| return self.callback_fn(pipeline, step_index, timestep, callback_kwargs) |
|
|
|
|
| class MultiPipelineCallbacks: |
|
|
| def __init__(self, callbacks: List[PipelineCallback]): |
| self.callbacks = callbacks |
|
|
| @property |
| def tensor_inputs(self) -> List[str]: |
| return [input for callback in self.callbacks for input in callback.tensor_inputs] |
|
|
| def __call__(self, pipeline, step_index, timestep, callback_kwargs) -> Dict[str, Any]: |
| """ |
| Calls all the callbacks in order with the given arguments and returns the final callback_kwargs. |
| """ |
| for callback in self.callbacks: |
| callback_kwargs = callback(pipeline, step_index, timestep, callback_kwargs) |
|
|
| return callback_kwargs |
|
|
| def seed_everything(seed): |
| torch.manual_seed(seed) |
| torch.cuda.manual_seed(seed) |
| random.seed(seed) |
| np.random.seed(seed) |
|
|
| def get_promptls(prompt_path): |
| with open(prompt_path) as f: |
| prompt_ls = json.load(f) |
| prompt_ls = [prompt['caption'].replace('/','_') for prompt in prompt_ls] |
| return prompt_ls |
|
|
|
|
|
|
| def warpped_feature(sample, step): |
| """ |
| sample: batch_size*dim*h*w, uncond: 0 - batch_size//2, cond: batch_size//2 - batch_size |
| step: timestep span |
| """ |
| bs, dim, h, w = sample.shape |
| |
| |
| |
| |
| if sample.shape[0] >= 2: |
| uncond_fea, cond_fea = sample.chunk(2, dim=0) |
| uncond_fea = uncond_fea.repeat(step,1,1,1) |
| cond_fea = cond_fea.repeat(step,1,1,1) |
| else: |
| uncond_fea = sample.repeat(step,1,1,1) |
| cond_fea = sample.repeat(step,1,1,1) |
| print("Warning: sample has only one chunk. Proceeding with only uncond_fea") |
| |
| return torch.cat([uncond_fea, cond_fea]) |
|
|
| def warpped_skip_feature(block_samples, step): |
| down_block_res_samples = [] |
| |
| for sample in block_samples: |
| sample_expand = warpped_feature(sample, step) |
| down_block_res_samples.append(sample_expand) |
| return tuple(down_block_res_samples) |
|
|
| def warpped_text_emb(text_emb, step): |
| """ |
| text_emb: batch_size*77*768, uncond: 0 - batch_size//2, cond: batch_size//2 - batch_size |
| step: timestep span |
| """ |
| bs, token_len, dim = text_emb.shape |
| if text_emb.shape[0] >= 2: |
| uncond_fea, cond_fea = text_emb.chunk(2) |
| uncond_fea = uncond_fea.repeat(step,1,1) |
| cond_fea = cond_fea.repeat(step,1,1) |
| else: |
| uncond_fea, cond_fea = text_emb, text_emb |
| uncond_fea = uncond_fea.repeat(step,1,1) |
| cond_fea = cond_fea.repeat(step,1,1) |
| return torch.cat([uncond_fea, cond_fea]) |
|
|
| def warpped_timestep(timesteps, bs): |
| """ |
| timestpes: list, such as [981, 961, 941] |
| """ |
| semi_bs = bs//2 |
| ts = [] |
| for timestep in timesteps: |
| timestep = timestep[None] |
| texp = timestep.expand(semi_bs) |
| ts.append(texp) |
| timesteps = torch.cat(ts) |
| return timesteps.repeat(2,1).reshape(-1) |
|
|
| def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): |
| """ |
| Rescale `noise_cfg` according to `guidance_rescale`. Based on findings of [Common Diffusion Noise Schedules and |
| Sample Steps are Flawed](https://arxiv.org/pdf/2305.08891.pdf). See Section 3.4 |
| """ |
| std_text = noise_pred_text.std(dim=list(range(1, noise_pred_text.ndim)), keepdim=True) |
| std_cfg = noise_cfg.std(dim=list(range(1, noise_cfg.ndim)), keepdim=True) |
| |
| noise_pred_rescaled = noise_cfg * (std_text / std_cfg) |
| |
| noise_cfg = guidance_rescale * noise_pred_rescaled + (1 - guidance_rescale) * noise_cfg |
| return noise_cfg |
|
|
| def register_parallel_pipeline_orig(pipe, mod='50ls4'): |
| def new_call(self): |
| @torch.no_grad() |
| def call( |
| prompt: Union[str, List[str]] = None, |
| prompt_2: Optional[Union[str, List[str]]] = None, |
| height: Optional[int] = None, |
| width: Optional[int] = None, |
| num_inference_steps: int = 50, |
| timesteps: List[int] = None, |
| sigmas: List[float] = None, |
| denoising_end: Optional[float] = None, |
| guidance_scale: float = 5.0, |
| negative_prompt: Optional[Union[str, List[str]]] = None, |
| negative_prompt_2: Optional[Union[str, List[str]]] = None, |
| num_images_per_prompt: Optional[int] = 1, |
| eta: float = 0.0, |
| generator: Optional[Union[torch.Generator, List[torch.Generator]]] = None, |
| latents: Optional[torch.Tensor] = None, |
| prompt_embeds: Optional[torch.Tensor] = None, |
| negative_prompt_embeds: Optional[torch.Tensor] = None, |
| pooled_prompt_embeds: Optional[torch.Tensor] = None, |
| negative_pooled_prompt_embeds: Optional[torch.Tensor] = None, |
| ip_adapter_image: Optional[PipelineImageInput] = None, |
| ip_adapter_image_embeds: Optional[List[torch.Tensor]] = None, |
| output_type: Optional[str] = "pil", |
| return_dict: bool = True, |
| cross_attention_kwargs: Optional[Dict[str, Any]] = None, |
| guidance_rescale: float = 0.0, |
| original_size: Optional[Tuple[int, int]] = None, |
| crops_coords_top_left: Tuple[int, int] = (0, 0), |
| target_size: Optional[Tuple[int, int]] = None, |
| negative_original_size: Optional[Tuple[int, int]] = None, |
| negative_crops_coords_top_left: Tuple[int, int] = (0, 0), |
| negative_target_size: Optional[Tuple[int, int]] = None, |
| clip_skip: Optional[int] = None, |
| callback_on_step_end: Optional[ |
| Union[Callable[[int, int, Dict], None], PipelineCallback, MultiPipelineCallbacks] |
| ] = None, |
| callback_on_step_end_tensor_inputs: List[str] = ["latents"], |
| **kwargs, |
| ): |
|
|
| callback = kwargs.pop("callback", None) |
| callback_steps = kwargs.pop("callback_steps", None) |
|
|
| if callback is not None: |
| deprecate( |
| "callback", |
| "1.0.0", |
| "Passing `callback` as an input argument to `__call__` is deprecated, consider use `callback_on_step_end`", |
| ) |
| if callback_steps is not None: |
| deprecate( |
| "callback_steps", |
| "1.0.0", |
| "Passing `callback_steps` as an input argument to `__call__` is deprecated, consider use `callback_on_step_end`", |
| ) |
|
|
| if isinstance(callback_on_step_end, (PipelineCallback, MultiPipelineCallbacks)): |
| callback_on_step_end_tensor_inputs = callback_on_step_end.tensor_inputs |
|
|
| |
| height = height or self.default_sample_size * self.vae_scale_factor |
| width = width or self.default_sample_size * self.vae_scale_factor |
| original_size = original_size or (height, width) |
| target_size = target_size or (height, width) |
|
|
| |
| self.check_inputs( |
| prompt, |
| prompt_2, |
| height, |
| width, |
| callback_steps, |
| negative_prompt, |
| negative_prompt_2, |
| prompt_embeds, |
| negative_prompt_embeds, |
| pooled_prompt_embeds, |
| negative_pooled_prompt_embeds, |
| ip_adapter_image, |
| ip_adapter_image_embeds, |
| callback_on_step_end_tensor_inputs, |
| ) |
|
|
| self._guidance_scale = guidance_scale |
| self._guidance_rescale = guidance_rescale |
| self._clip_skip = clip_skip |
| self._cross_attention_kwargs = cross_attention_kwargs |
| self._denoising_end = denoising_end |
| self._interrupt = False |
|
|
| |
| if prompt is not None and isinstance(prompt, str): |
| batch_size = 1 |
| elif prompt is not None and isinstance(prompt, list): |
| batch_size = len(prompt) |
| else: |
| batch_size = prompt_embeds.shape[0] |
|
|
| device = self._execution_device |
|
|
| |
| lora_scale = ( |
| self.cross_attention_kwargs.get("scale", None) if self.cross_attention_kwargs is not None else None |
| ) |
|
|
| ( |
| prompt_embeds, |
| negative_prompt_embeds, |
| pooled_prompt_embeds, |
| negative_pooled_prompt_embeds, |
| ) = self.encode_prompt( |
| prompt=prompt, |
| prompt_2=prompt_2, |
| device=device, |
| num_images_per_prompt=num_images_per_prompt, |
| do_classifier_free_guidance=self.do_classifier_free_guidance, |
| negative_prompt=negative_prompt, |
| negative_prompt_2=negative_prompt_2, |
| prompt_embeds=prompt_embeds, |
| negative_prompt_embeds=negative_prompt_embeds, |
| pooled_prompt_embeds=pooled_prompt_embeds, |
| negative_pooled_prompt_embeds=negative_pooled_prompt_embeds, |
| lora_scale=lora_scale, |
| clip_skip=self.clip_skip, |
| ) |
|
|
| if self.do_classifier_free_guidance: |
| prompt_embeds = torch.cat([negative_prompt_embeds, prompt_embeds], dim=0) |
| pooled_prompt_embeds = torch.cat([negative_pooled_prompt_embeds, pooled_prompt_embeds], dim=0) |
|
|
|
|
| |
| self.scheduler.set_timesteps(num_inference_steps, device=device) |
| timesteps = self.scheduler.timesteps |
|
|
| |
| num_channels_latents = self.unet.config.in_channels |
| latents = self.prepare_latents( |
| batch_size * num_images_per_prompt, |
| num_channels_latents, |
| height, |
| width, |
| prompt_embeds.dtype, |
| device, |
| generator, |
| latents, |
| ) |
|
|
| |
| extra_step_kwargs = self.prepare_extra_step_kwargs(generator, eta) |
|
|
| |
| timestep_cond = None |
| if self.unet.config.time_cond_proj_dim is not None: |
| guidance_scale_tensor = torch.tensor(self.guidance_scale - 1).repeat(batch_size * num_images_per_prompt) |
| timestep_cond = self.get_guidance_scale_embedding( |
| guidance_scale_tensor, embedding_dim=self.unet.config.time_cond_proj_dim |
| ).to(device=device, dtype=latents.dtype) |
|
|
| |
| init_latents = latents.detach().clone() |
| all_steps = len(self.scheduler.timesteps) |
| curr_step = 0 |
| if mod == '50ls': |
| cond = lambda timestep: timestep in [0, 1, 2, 3, 5, 10, 15, 25, 35] |
| elif isinstance(mod, int): |
| cond = lambda timestep: timestep % mod == 0 |
| else: |
| raise Exception("Currently not supported, But you can modify the code to customize the keytime") |
|
|
| |
| add_time_ids = self._get_add_time_ids( |
| original_size, |
| crops_coords_top_left, |
| target_size, |
| dtype=prompt_embeds.dtype, |
| text_encoder_projection_dim=pooled_prompt_embeds.shape[-1] if self.text_encoder_2 is None else self.text_encoder_2.config.projection_dim, |
| ) |
| if negative_original_size is not None and negative_target_size is not None: |
| negative_add_time_ids = self._get_add_time_ids( |
| negative_original_size, |
| negative_crops_coords_top_left, |
| negative_target_size, |
| dtype=prompt_embeds.dtype, |
| text_encoder_projection_dim=pooled_prompt_embeds.shape[-1] if self.text_encoder_2 is None else self.text_encoder_2.config.projection_dim, |
| ) |
| else: |
| negative_add_time_ids = add_time_ids |
| |
| if self.do_classifier_free_guidance: |
| add_time_ids = torch.cat([negative_add_time_ids, add_time_ids], dim=0) |
|
|
| add_text_embeds = pooled_prompt_embeds.to(device) |
| add_time_ids = add_time_ids.to(device).repeat(batch_size * num_images_per_prompt, 1) |
|
|
|
|
| while curr_step < all_steps: |
| register_time(self.unet, curr_step) |
| time_ls = [timesteps[curr_step]] |
| curr_step += 1 |
| while not cond(curr_step) and curr_step < all_steps: |
| time_ls.append(timesteps[curr_step]) |
| curr_step += 1 |
|
|
| latent_model_input = torch.cat([latents] * 2) if self.do_classifier_free_guidance else latents |
|
|
| |
| added_cond_kwargs = {"text_embeds": add_text_embeds, "time_ids": add_time_ids} |
| if ip_adapter_image is not None or ip_adapter_image_embeds is not None: |
| image_embeds = self.prepare_ip_adapter_image_embeds( |
| ip_adapter_image, |
| ip_adapter_image_embeds, |
| device, |
| batch_size * num_images_per_prompt, |
| self.do_classifier_free_guidance, |
| ) |
| added_cond_kwargs["image_embeds"] = image_embeds |
|
|
| noise_pred = self.unet( |
| latent_model_input, |
| time_ls, |
| encoder_hidden_states=prompt_embeds, |
| timestep_cond=timestep_cond, |
| cross_attention_kwargs=self.cross_attention_kwargs, |
| added_cond_kwargs=added_cond_kwargs, |
| return_dict=False, |
| )[0] |
|
|
| if self.do_classifier_free_guidance: |
| noise_pred_uncond, noise_pred_text = noise_pred.chunk(2) |
| noise_pred = noise_pred_uncond + self.guidance_scale * (noise_pred_text - noise_pred_uncond) |
|
|
| if self.do_classifier_free_guidance and self.guidance_rescale > 0.0: |
| noise_pred = rescale_noise_cfg(noise_pred, noise_pred_text, guidance_rescale=self.guidance_rescale) |
|
|
| bs = noise_pred.shape[0] |
| bs_perstep = bs // len(time_ls) |
| denoised_latent = latents |
| for i, timestep in enumerate(time_ls): |
| if timestep / 1000 < 0.5: |
| denoised_latent = denoised_latent + 0.003 * init_latents |
| curr_noise = noise_pred[i * bs_perstep : (i + 1) * bs_perstep] |
| denoised_latent = self.scheduler.step( |
| curr_noise, timestep, denoised_latent, **extra_step_kwargs, return_dict=False |
| )[0] |
| latents = denoised_latent |
|
|
| |
| if not output_type == "latent": |
| needs_upcasting = self.vae.dtype == torch.float16 and self.vae.config.force_upcast |
| if needs_upcasting: |
| self.upcast_vae() |
| latents = latents.to(next(iter(self.vae.post_quant_conv.parameters())).dtype) |
| elif latents.dtype != self.vae.dtype: |
| if torch.backends.mps.is_available(): |
| self.vae = self.vae.to(latents.dtype) |
|
|
| has_latents_mean = hasattr(self.vae.config, "latents_mean") and self.vae.config.latents_mean is not None |
| has_latents_std = hasattr(self.vae.config, "latents_std") and self.vae.config.latents_std is not None |
| if has_latents_mean and has_latents_std: |
| latents_mean = ( |
| torch.tensor(self.vae.config.latents_mean).view(1, 4, 1, 1).to(latents.device, latents.dtype) |
| ) |
| latents_std = ( |
| torch.tensor(self.vae.config.latents_std).view(1, 4, 1, 1).to(latents.device, latents.dtype) |
| ) |
| latents = latents * latents_std / self.vae.config.scaling_factor + latents_mean |
| else: |
| latents = latents / self.vae.config.scaling_factor |
|
|
| image = self.vae.decode(latents, return_dict=False)[0] |
|
|
| if needs_upcasting: |
| self.vae.to(dtype=torch.float16) |
| else: |
| image = latents |
|
|
| if not output_type == "latent": |
| if self.watermark is not None: |
| image = self.watermark.apply_watermark(image) |
|
|
| image = self.image_processor.postprocess(image, output_type=output_type) |
|
|
| self.maybe_free_model_hooks() |
|
|
| if not return_dict: |
| return (image,) |
|
|
| return StableDiffusionXLPipelineOutput(images=image) |
|
|
| return call |
|
|
| pipe.call = new_call(pipe) |
|
|
| def register_faster_forward(model, mod = '50ls4'): |
| def faster_forward(self): |
| def forward( |
| sample: torch.FloatTensor, |
| timestep: Union[torch.Tensor, float, int], |
| encoder_hidden_states: torch.Tensor, |
| class_labels: Optional[torch.Tensor] = None, |
| timestep_cond: Optional[torch.Tensor] = None, |
| attention_mask: Optional[torch.Tensor] = None, |
| cross_attention_kwargs: Optional[Dict[str, Any]] = None, |
| added_cond_kwargs: Optional[Dict[str, torch.Tensor]] = None, |
| down_block_additional_residuals: Optional[Tuple[torch.Tensor]] = None, |
| mid_block_additional_residual: Optional[torch.Tensor] = None, |
| down_intrablock_additional_residuals: Optional[Tuple[torch.Tensor]] = None, |
| return_dict: bool = True, |
| ) -> Union[UNet2DConditionOutput, Tuple]: |
| r""" |
| Args: |
| sample (`torch.FloatTensor`): (batch, channel, height, width) noisy inputs tensor |
| timestep (`torch.FloatTensor` or `float` or `int`): (batch) timesteps |
| encoder_hidden_states (`torch.FloatTensor`): (batch, sequence_length, feature_dim) encoder hidden states |
| return_dict (`bool`, *optional*, defaults to `True`): |
| Whether or not to return a [`models.unet_2d_condition.UNet2DConditionOutput`] instead of a plain tuple. |
| cross_attention_kwargs (`dict`, *optional*): |
| A kwargs dictionary that if specified is passed along to the `AttentionProcessor` as defined under |
| `self.processor` in |
| [diffusers.cross_attention](https://github.com/huggingface/diffusers/blob/main/src/diffusers/models/cross_attention.py). |
| |
| Returns: |
| [`~models.unet_2d_condition.UNet2DConditionOutput`] or `tuple`: |
| [`~models.unet_2d_condition.UNet2DConditionOutput`] if `return_dict` is True, otherwise a `tuple`. When |
| returning a tuple, the first element is the sample tensor. |
| """ |
| |
| |
| |
| |
| default_overall_up_factor = 2**self.num_upsamplers |
|
|
| |
| forward_upsample_size = False |
| upsample_size = None |
|
|
| if any(s % default_overall_up_factor != 0 for s in sample.shape[-2:]): |
| print("Forward upsample size to force interpolation output size.") |
| forward_upsample_size = True |
|
|
| |
| if attention_mask is not None: |
| attention_mask = (1 - attention_mask.to(sample.dtype)) * -10000.0 |
| attention_mask = attention_mask.unsqueeze(1) |
|
|
| |
| if self.config.center_input_sample: |
| sample = 2 * sample - 1.0 |
|
|
| |
| step = 1 |
| |
| |
| |
| |
| |
| |
| if not torch.is_tensor(timesteps) and (not isinstance(timesteps,list)): |
| |
| |
| is_mps = sample.device.type == "mps" |
| if isinstance(timestep, float): |
| dtype = torch.float32 if is_mps else torch.float64 |
| else: |
| dtype = torch.int32 if is_mps else torch.int64 |
| timesteps = torch.tensor([timesteps], dtype=dtype, device=sample.device) |
| elif (not isinstance(timesteps,list)) and len(timesteps.shape) == 0: |
| timesteps = timesteps[None].to(sample.device) |
| |
| if (not isinstance(timesteps,list)) and len(timesteps.shape) == 1: |
| |
| timesteps = timesteps.expand(sample.shape[0]) |
| elif isinstance(timesteps, list): |
| |
| timesteps = warpped_timestep(timesteps, sample.shape[0]).to(sample.device) |
| t_emb = self.time_proj(timesteps) |
|
|
| |
| |
| |
| t_emb = t_emb.to(dtype=self.dtype) |
|
|
| emb = self.time_embedding(t_emb, timestep_cond) |
|
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| if self.class_embedding is not None: |
| if class_labels is None: |
| raise ValueError("class_labels should be provided when num_class_embeds > 0") |
|
|
| if self.config.class_embed_type == "timestep": |
| class_labels = self.time_proj(class_labels) |
|
|
| |
| |
| class_labels = class_labels.to(dtype=sample.dtype) |
|
|
| class_emb = self.class_embedding(class_labels).to(dtype=self.dtype) |
|
|
| if self.config.class_embeddings_concat: |
| emb = torch.cat([emb, class_emb], dim=-1) |
| else: |
| emb = emb + class_emb |
|
|
| if self.config.addition_embed_type == "text": |
| aug_emb = self.add_embedding(encoder_hidden_states) |
| emb = emb + aug_emb |
|
|
| |
|
|
| if self.time_embed_act is not None: |
| emb = self.time_embed_act(emb) |
|
|
| if self.encoder_hid_proj is not None: |
| print("Encoder hid projections are not none") |
| encoder_hidden_states = self.encoder_hid_proj(encoder_hidden_states) |
|
|
| |
| order = self.order |
| |
| ipow = int(np.sqrt(9 + 8*order)) |
| cond = order in [0, 1, 2, 3, 5, 10, 15, 25, 35] |
| if isinstance(mod, int): |
| cond = order % mod == 0 |
| elif mod == "pro": |
| cond = ipow * ipow == (9 + 8 * order) |
| elif mod == "50ls": |
| cond = order in [0, 1, 2, 3, 5, 10, 15, 25, 35] |
| elif mod == "50ls2": |
| cond = order in [0, 10, 11, 12, 15, 20, 25, 30,35,45] |
| elif mod == "50ls3": |
| cond = order in [0, 20, 25, 30,35,45,46,47,48,49] |
| elif mod == "50ls4": |
| cond = order in [0, 9, 13, 14, 15, 28, 29, 32, 36,45] |
| elif mod == "100ls": |
| cond = order > 85 or order < 10 or order % 5 == 0 |
| elif mod == "75ls": |
| cond = order > 65 or order < 10 or order % 5 == 0 |
| elif mod == "s2": |
| cond = order < 20 or order > 40 or order % 2 == 0 |
|
|
| if cond: |
| |
| |
| sample = self.conv_in(sample) |
|
|
| |
| down_block_res_samples = (sample,) |
| print(f"Shape of sample at the start of downsample: {sample.shape}") |
| for i, downsample_block in enumerate(self.down_blocks): |
| print(f"\n--- Downsample Block {i+1} ---") |
| print(f" Input shape: {sample.shape}") |
| if hasattr(downsample_block, "has_cross_attention") and downsample_block.has_cross_attention: |
| print("has cross attention") |
| additional_residuals = {} |
| if down_intrablock_additional_residuals is not None and len(down_intrablock_additional_residuals) > 0: |
| additional_residuals["additional_residuals"] = down_intrablock_additional_residuals.pop(0) |
| print(f" Additional Residuals shape: {additional_residuals['additional_residuals'].shape}") |
| |
| sample, res_samples = downsample_block( |
| hidden_states=sample, |
| temb=emb, |
| encoder_hidden_states=encoder_hidden_states, |
| attention_mask=attention_mask, |
| cross_attention_kwargs=cross_attention_kwargs, |
| **additional_residuals |
| ) |
| else: |
| sample, res_samples = downsample_block(hidden_states=sample, temb=emb) |
| |
| |
| print(f" Output shape: {sample.shape}") |
| for j, res_sample in enumerate(res_samples): |
| print(f" Res sample {j+1} shape: {res_sample.shape}") |
| down_block_res_samples += res_samples |
|
|
| |
|
|
| if down_block_additional_residuals is not None: |
| new_down_block_res_samples = () |
|
|
| for down_block_res_sample, down_block_additional_residual in zip( |
| down_block_res_samples, down_block_additional_residuals |
| ): |
| down_block_res_sample = down_block_res_sample + down_block_additional_residual |
| new_down_block_res_samples += (down_block_res_sample,) |
|
|
| down_block_res_samples = new_down_block_res_samples |
|
|
| |
| if down_block_additional_residuals is not None: |
| new_down_block_res_samples = () |
| for down_block_res_sample, down_block_additional_residual in zip( |
| down_block_res_samples, down_block_additional_residuals |
| ): |
| down_block_res_sample = down_block_res_sample + down_block_additional_residual |
| new_down_block_res_samples += (down_block_res_sample,) |
| down_block_res_samples = new_down_block_res_samples |
| |
| |
| if self.mid_block is not None: |
| if hasattr(self.mid_block, "has_cross_attention") and self.mid_block.has_cross_attention: |
| sample = self.mid_block( |
| sample, |
| emb, |
| encoder_hidden_states=encoder_hidden_states, |
| attention_mask=attention_mask, |
| cross_attention_kwargs=cross_attention_kwargs, |
| ) |
| else: |
| sample = self.mid_block(sample, emb) |
| |
| if down_intrablock_additional_residuals is not None and len(down_intrablock_additional_residuals) > 0 and sample.shape == down_intrablock_additional_residuals[0].shape: |
| sample += down_intrablock_additional_residuals.pop(0) |
|
|
| if mid_block_additional_residual is not None: |
| sample = sample + mid_block_additional_residual |
|
|
| |
| |
| setattr(self, 'skip_feature', deepcopy(down_block_res_samples)) |
| setattr(self, 'toup_feature', sample.detach().clone()) |
| |
|
|
|
|
|
|
| |
| if isinstance(timestep, list): |
| |
| timesteps = warpped_timestep(timestep, sample.shape[0]).to(sample.device) |
| t_emb = self.time_proj(timesteps) |
|
|
| |
| |
| |
| t_emb = t_emb.to(dtype=self.dtype) |
|
|
| emb = self.time_embedding(t_emb, timestep_cond) |
|
|
| down_block_res_samples = warpped_skip_feature(down_block_res_samples, step) |
| sample = warpped_feature(sample, step) |
| encoder_hidden_states = warpped_text_emb(encoder_hidden_states, step) |
| |
| |
| else: |
| down_block_res_samples = self.skip_feature |
| sample = self.toup_feature |
|
|
| |
| down_block_res_samples = warpped_skip_feature(down_block_res_samples, step) |
| sample = warpped_feature(sample, step) |
| encoder_hidden_states = warpped_text_emb(encoder_hidden_states, step) |
| |
|
|
| |
| for i, upsample_block in enumerate(self.up_blocks): |
| is_final_block = i == len(self.up_blocks) - 1 |
|
|
| res_samples = down_block_res_samples[-len(upsample_block.resnets) :] |
| down_block_res_samples = down_block_res_samples[: -len(upsample_block.resnets)] |
|
|
| |
| |
| if not is_final_block and forward_upsample_size: |
| upsample_size = down_block_res_samples[-1].shape[2:] |
|
|
| if hasattr(upsample_block, "has_cross_attention") and upsample_block.has_cross_attention: |
| sample = upsample_block( |
| hidden_states=sample, |
| temb=emb, |
| res_hidden_states_tuple=res_samples, |
| encoder_hidden_states=encoder_hidden_states, |
| cross_attention_kwargs=cross_attention_kwargs, |
| upsample_size=upsample_size, |
| attention_mask=attention_mask, |
| ) |
| else: |
| sample = upsample_block( |
| hidden_states=sample, temb=emb, res_hidden_states_tuple=res_samples, upsample_size=upsample_size |
| ) |
|
|
| |
| if self.conv_norm_out: |
| sample = self.conv_norm_out(sample) |
| sample = self.conv_act(sample) |
| sample = self.conv_out(sample) |
|
|
| if not return_dict: |
| return (sample,) |
|
|
| return UNet2DConditionOutput(sample=sample) |
| return forward |
| if model.__class__.__name__ == 'UNet2DConditionModel': |
| model.forward = faster_forward(model) |
|
|
| def register_faster_orig_forward(model, mod = '50ls4'): |
| def faster_orig_forward(self): |
| def forward( |
| sample: torch.FloatTensor, |
| timestep: Union[torch.Tensor, float, int], |
| encoder_hidden_states: torch.Tensor, |
| class_labels: Optional[torch.Tensor] = None, |
| timestep_cond: Optional[torch.Tensor] = None, |
| attention_mask: Optional[torch.Tensor] = None, |
| cross_attention_kwargs: Optional[Dict[str, Any]] = None, |
| added_cond_kwargs: Optional[Dict[str, torch.Tensor]] = None, |
| down_block_additional_residuals: Optional[Tuple[torch.Tensor]] = None, |
| mid_block_additional_residual: Optional[torch.Tensor] = None, |
| down_intrablock_additional_residuals: Optional[Tuple[torch.Tensor]] = None, |
| encoder_attention_mask: Optional[torch.Tensor] = None, |
| return_dict: bool = True, |
| ) -> Union[UNet2DConditionOutput, Tuple]: |
| |
| default_overall_up_factor = 2**self.num_upsamplers |
| |
| |
| forward_upsample_size = False |
| upsample_size = None |
| |
| for dim in sample.shape[-2:]: |
| if dim % default_overall_up_factor != 0: |
| |
| forward_upsample_size = True |
| break |
| |
| if attention_mask is not None: |
| attention_mask = (1 - attention_mask.to(sample.dtype)) * -10000.0 |
| attention_mask = attention_mask.unsqueeze(1) |
| |
| |
| if encoder_attention_mask is not None: |
| encoder_attention_mask = (1 - encoder_attention_mask.to(sample.dtype)) * -10000.0 |
| encoder_attention_mask = encoder_attention_mask.unsqueeze(1) |
| |
| |
| if self.config.center_input_sample: |
| sample = 2 * sample - 1.0 |
| |
| |
| timesteps = timestep |
| if not torch.is_tensor(timesteps): |
| is_mps = sample.device.type == "mps" |
| if isinstance(timestep, float): |
| dtype = torch.float32 if is_mps else torch.float64 |
| else: |
| dtype = torch.int32 if is_mps else torch.int64 |
| timesteps = torch.tensor([timesteps], dtype=dtype, device=sample.device) |
| elif len(timesteps.shape) == 0: |
| timesteps = timesteps[None].to(sample.device) |
| |
| |
| timesteps = timesteps.expand(sample.shape[0]) |
| |
| t_emb = self.time_proj(timesteps) |
| |
| t_emb = t_emb.to(dtype=sample.dtype) |
| |
| emb = self.time_embedding(t_emb, timestep_cond) |
| aug_emb = None |
| |
| if self.class_embedding is not None: |
| if class_labels is None: |
| raise ValueError("class_labels should be provided when num_class_embeds > 0") |
| |
| if self.config.class_embed_type == "timestep": |
| class_labels = self.time_proj(class_labels) |
| class_labels = class_labels.to(dtype=sample.dtype) |
| |
| class_emb = self.class_embedding(class_labels).to(dtype=sample.dtype) |
| |
| if self.config.class_embeddings_concat: |
| emb = torch.cat([emb, class_emb], dim=-1) |
| else: |
| emb = emb + class_emb |
| |
| if self.config.addition_embed_type == "text": |
| aug_emb = self.add_embedding(encoder_hidden_states) |
| elif self.config.addition_embed_type == "text_image": |
| |
| if "image_embeds" not in added_cond_kwargs: |
| raise ValueError( |
| f"{self.__class__} has the config param `addition_embed_type` set to 'text_image' which requires the keyword argument `image_embeds` to be passed in `added_cond_kwargs`" |
| ) |
| |
| image_embs = added_cond_kwargs.get("image_embeds") |
| text_embs = added_cond_kwargs.get("text_embeds", encoder_hidden_states) |
| aug_emb = self.add_embedding(text_embs, image_embs) |
| elif self.config.addition_embed_type == "text_time": |
| |
| if "text_embeds" not in added_cond_kwargs: |
| raise ValueError( |
| f"{self.__class__} has the config param `addition_embed_type` set to 'text_time' which requires the keyword argument `text_embeds` to be passed in `added_cond_kwargs`" |
| ) |
| text_embeds = added_cond_kwargs.get("text_embeds") |
| if "time_ids" not in added_cond_kwargs: |
| raise ValueError( |
| f"{self.__class__} has the config param `addition_embed_type` set to 'text_time' which requires the keyword argument `time_ids` to be passed in `added_cond_kwargs`" |
| ) |
| time_ids = added_cond_kwargs.get("time_ids") |
| time_embeds = self.add_time_proj(time_ids.flatten()) |
| time_embeds = time_embeds.reshape((text_embeds.shape[0], -1)) |
| add_embeds = torch.concat([text_embeds, time_embeds], dim=-1) |
| add_embeds = add_embeds.to(emb.dtype) |
| aug_emb = self.add_embedding(add_embeds) |
| elif self.config.addition_embed_type == "image": |
| |
| if "image_embeds" not in added_cond_kwargs: |
| raise ValueError( |
| f"{self.__class__} has the config param `addition_embed_type` set to 'image' which requires the keyword argument `image_embeds` to be passed in `added_cond_kwargs`" |
| ) |
| image_embs = added_cond_kwargs.get("image_embeds") |
| aug_emb = self.add_embedding(image_embs) |
| elif self.config.addition_embed_type == "image_hint": |
| |
| if "image_embeds" not in added_cond_kwargs or "hint" not in added_cond_kwargs: |
| raise ValueError( |
| f"{self.__class__} has the config param `addition_embed_type` set to 'image_hint' which requires the keyword arguments `image_embeds` and `hint` to be passed in `added_cond_kwargs`" |
| ) |
| image_embs = added_cond_kwargs.get("image_embeds") |
| hint = added_cond_kwargs.get("hint") |
| aug_emb, hint = self.add_embedding(image_embs, hint) |
| sample = torch.cat([sample, hint], dim=1) |
| |
| emb = emb + aug_emb if aug_emb is not None else emb |
| |
| if self.time_embed_act is not None: |
| emb = self.time_embed_act(emb) |
| |
| if self.encoder_hid_proj is not None and self.config.encoder_hid_dim_type == "text_proj": |
| encoder_hidden_states = self.encoder_hid_proj(encoder_hidden_states) |
| elif self.encoder_hid_proj is not None and self.config.encoder_hid_dim_type == "text_image_proj": |
| |
| if "image_embeds" not in added_cond_kwargs: |
| raise ValueError( |
| f"{self.__class__} has the config param `encoder_hid_dim_type` set to 'text_image_proj' which requires the keyword argument `image_embeds` to be passed in `added_conditions`" |
| ) |
| |
| image_embeds = added_cond_kwargs.get("image_embeds") |
| encoder_hidden_states = self.encoder_hid_proj(encoder_hidden_states, image_embeds) |
| elif self.encoder_hid_proj is not None and self.config.encoder_hid_dim_type == "image_proj": |
| |
| if "image_embeds" not in added_cond_kwargs: |
| raise ValueError( |
| f"{self.__class__} has the config param `encoder_hid_dim_type` set to 'image_proj' which requires the keyword argument `image_embeds` to be passed in `added_conditions`" |
| ) |
| image_embeds = added_cond_kwargs.get("image_embeds") |
| encoder_hidden_states = self.encoder_hid_proj(image_embeds) |
| elif self.encoder_hid_proj is not None and self.config.encoder_hid_dim_type == "ip_image_proj": |
| if "image_embeds" not in added_cond_kwargs: |
| raise ValueError( |
| f"{self.__class__} has the config param `encoder_hid_dim_type` set to 'ip_image_proj' which requires the keyword argument `image_embeds` to be passed in `added_conditions`" |
| ) |
| image_embeds = added_cond_kwargs.get("image_embeds") |
| image_embeds = self.encoder_hid_proj(image_embeds).to(encoder_hidden_states.dtype) |
| encoder_hidden_states = torch.cat([encoder_hidden_states, image_embeds], dim=1) |
|
|
| |
| order = model.order |
| |
| ipow = int(np.sqrt(9 + 8*order)) |
| cond = order in [0, 1, 2, 3, 5, 10, 15, 25, 35] |
| if isinstance(mod, int): |
| cond = order % mod == 0 |
| elif mod == "pro": |
| cond = ipow * ipow == (9 + 8 * order) |
| elif mod == "50ls": |
| cond = order in [0, 1, 2, 3, 5, 10, 15, 25, 35] |
| elif mod == "50ls2": |
| cond = order in [0, 10, 11, 12, 15, 20, 25, 30,35,45] |
| elif mod == "50ls3": |
| cond = order in [0, 20, 25, 30,35,45,46,47,48,49] |
| elif mod == "50ls4": |
| cond = order in [0, 9, 13, 14, 15, 28, 29, 32, 36,45] |
| elif mod == "100ls": |
| cond = order > 85 or order < 10 or order % 5 == 0 |
| elif mod == "75ls": |
| cond = order > 65 or order < 10 or order % 5 == 0 |
| elif mod == "s2": |
| cond = order < 20 or order > 40 or order % 2 == 0 |
|
|
| if cond: |
| |
| sample = self.conv_in(sample) |
| |
| |
| if cross_attention_kwargs is not None and cross_attention_kwargs.get("gligen", None) is not None: |
| cross_attention_kwargs = cross_attention_kwargs.copy() |
| gligen_args = cross_attention_kwargs.pop("gligen") |
| cross_attention_kwargs["gligen"] = {"objs": self.position_net(**gligen_args)} |
| |
| |
| lora_scale = cross_attention_kwargs.get("scale", 1.0) if cross_attention_kwargs is not None else 1.0 |
| if USE_PEFT_BACKEND: |
| |
| scale_lora_layers(self, lora_scale) |
| |
| is_controlnet = mid_block_additional_residual is not None and down_block_additional_residuals is not None |
| |
| is_adapter = down_intrablock_additional_residuals is not None |
| |
| |
| |
| if not is_adapter and mid_block_additional_residual is None and down_block_additional_residuals is not None: |
| deprecate( |
| "T2I should not use down_block_additional_residuals", |
| "1.3.0", |
| "Passing intrablock residual connections with `down_block_additional_residuals` is deprecated \ |
| and will be removed in diffusers 1.3.0. `down_block_additional_residuals` should only be used \ |
| for ControlNet. Please make sure use `down_intrablock_additional_residuals` instead. ", |
| standard_warn=False, |
| ) |
| down_intrablock_additional_residuals = down_block_additional_residuals |
| is_adapter = True |
| |
| down_block_res_samples = (sample,) |
| for downsample_block in self.down_blocks: |
| if hasattr(downsample_block, "has_cross_attention") and downsample_block.has_cross_attention: |
| |
| additional_residuals = {} |
| if is_adapter and len(down_intrablock_additional_residuals) > 0: |
| additional_residuals["additional_residuals"] = down_intrablock_additional_residuals.pop(0) |
| |
| sample, res_samples = downsample_block( |
| hidden_states=sample, |
| temb=emb, |
| encoder_hidden_states=encoder_hidden_states, |
| attention_mask=attention_mask, |
| cross_attention_kwargs=cross_attention_kwargs, |
| encoder_attention_mask=encoder_attention_mask, |
| **additional_residuals, |
| ) |
| else: |
| sample, res_samples = downsample_block(hidden_states=sample, temb=emb, scale=lora_scale) |
| if is_adapter and len(down_intrablock_additional_residuals) > 0: |
| sample += down_intrablock_additional_residuals.pop(0) |
| |
| down_block_res_samples += res_samples |
| |
| if is_controlnet: |
| new_down_block_res_samples = () |
| |
| for down_block_res_sample, down_block_additional_residual in zip( |
| down_block_res_samples, down_block_additional_residuals |
| ): |
| down_block_res_sample = down_block_res_sample + down_block_additional_residual |
| new_down_block_res_samples = new_down_block_res_samples + (down_block_res_sample,) |
| |
| down_block_res_samples = new_down_block_res_samples |
| |
| |
| if self.mid_block is not None: |
| if hasattr(self.mid_block, "has_cross_attention") and self.mid_block.has_cross_attention: |
| sample = self.mid_block( |
| sample, |
| emb, |
| encoder_hidden_states=encoder_hidden_states, |
| attention_mask=attention_mask, |
| cross_attention_kwargs=cross_attention_kwargs, |
| encoder_attention_mask=encoder_attention_mask, |
| ) |
| else: |
| sample = self.mid_block(sample, emb) |
| |
| |
| if ( |
| is_adapter |
| and len(down_intrablock_additional_residuals) > 0 |
| and sample.shape == down_intrablock_additional_residuals[0].shape |
| ): |
| sample += down_intrablock_additional_residuals.pop(0) |
| |
| if is_controlnet: |
| sample = sample + mid_block_additional_residual |
|
|
| |
| |
| setattr(self, 'skip_feature', deepcopy(down_block_res_samples)) |
| setattr(self, 'toup_feature', sample.detach().clone()) |
| |
|
|
|
|
|
|
| |
| if isinstance(timestep, list): |
| step = len(timestep) |
| |
| timesteps = warpped_timestep(timestep, sample.shape[0]).to(sample.device) |
| t_emb = self.time_proj(timesteps) |
|
|
| |
| |
| |
| t_emb = t_emb.to(dtype=self.dtype) |
|
|
| emb = self.time_embedding(t_emb, timestep_cond) |
| else: |
| step = 1 |
| down_block_res_samples = warpped_skip_feature(down_block_res_samples, step) |
| sample = warpped_feature(sample, step) |
| encoder_hidden_states = warpped_text_emb(encoder_hidden_states, step) |
| |
|
|
| else: |
| step = 1 |
| down_block_res_samples = self.skip_feature |
| sample = self.toup_feature |
|
|
| |
| down_block_res_samples = warpped_skip_feature(down_block_res_samples, step) |
| sample = warpped_feature(sample, step) |
| encoder_hidden_states = warpped_text_emb(encoder_hidden_states, step) |
| |
| |
| |
| for i, upsample_block in enumerate(self.up_blocks): |
| is_final_block = i == len(self.up_blocks) - 1 |
| |
| res_samples = down_block_res_samples[-len(upsample_block.resnets) :] |
| down_block_res_samples = down_block_res_samples[: -len(upsample_block.resnets)] |
| |
| |
| |
| if not is_final_block and forward_upsample_size: |
| upsample_size = down_block_res_samples[-1].shape[2:] |
| |
| if hasattr(upsample_block, "has_cross_attention") and upsample_block.has_cross_attention: |
| sample = upsample_block( |
| hidden_states=sample, |
| temb=emb, |
| res_hidden_states_tuple=res_samples, |
| encoder_hidden_states=encoder_hidden_states, |
| cross_attention_kwargs=cross_attention_kwargs, |
| upsample_size=upsample_size, |
| attention_mask=attention_mask, |
| encoder_attention_mask=encoder_attention_mask, |
| ) |
| else: |
| lora_scale = cross_attention_kwargs.get("scale", 1.0) if cross_attention_kwargs is not None else 1.0 |
| sample = upsample_block( |
| hidden_states=sample, |
| temb=emb, |
| res_hidden_states_tuple=res_samples, |
| upsample_size=upsample_size, |
| scale=lora_scale, |
| ) |
| |
| |
| if self.conv_norm_out: |
| sample = self.conv_norm_out(sample) |
| sample = self.conv_act(sample) |
| sample = self.conv_out(sample) |
| |
| if USE_PEFT_BACKEND: |
| |
| unscale_lora_layers(self, lora_scale) |
| |
| if not return_dict: |
| return (sample,) |
| |
| return UNet2DConditionOutput(sample=sample) |
| return forward |
| if model.__class__.__name__ == 'UNet2DConditionModel': |
| model.forward = faster_orig_forward(model) |
|
|
|
|
| def register_time(unet, t): |
| setattr(unet, 'order', t) |
|
|
| |