| | import inspect |
| | from typing import Callable, Dict, List, Optional, Union |
| |
|
| | import numpy as np |
| | import torch |
| | from transformers import PreTrainedModel, PreTrainedTokenizerFast |
| |
|
| | from diffusers.callbacks import MultiPipelineCallbacks, PipelineCallback |
| | from diffusers.models import AutoencoderKLWan, CosmosTransformer3DModel |
| | from diffusers.schedulers import FlowMatchEulerDiscreteScheduler |
| | from diffusers.utils import logging |
| | from diffusers.utils.torch_utils import randn_tensor |
| | from diffusers.video_processor import VideoProcessor |
| | from diffusers.pipelines.pipeline_utils import DiffusionPipeline |
| | from diffusers.pipelines.cosmos.pipeline_output import CosmosImagePipelineOutput |
| |
|
| | logger = logging.get_logger(__name__) |
| |
|
| |
|
| | def retrieve_timesteps(scheduler, num_inference_steps=None, device=None, timesteps=None, sigmas=None, **kwargs): |
| | if timesteps is not None and sigmas is not None: |
| | raise ValueError("Only one of `timesteps` or `sigmas` can be passed.") |
| | if timesteps is not None: |
| | scheduler.set_timesteps(timesteps=timesteps, device=device, **kwargs) |
| | timesteps = scheduler.timesteps |
| | num_inference_steps = len(timesteps) |
| | elif sigmas is not None: |
| | scheduler.set_timesteps(sigmas=sigmas, device=device, **kwargs) |
| | timesteps = scheduler.timesteps |
| | num_inference_steps = len(timesteps) |
| | else: |
| | scheduler.set_timesteps(num_inference_steps, device=device, **kwargs) |
| | timesteps = scheduler.timesteps |
| | return timesteps, num_inference_steps |
| |
|
| |
|
| | class AnimaTextToImagePipeline(DiffusionPipeline): |
| | """Pipeline for text-to-image generation using the Anima model. |
| | |
| | Anima uses a Cosmos Predict2 backbone with a Qwen3 text encoder and an LLM adapter |
| | that cross-attends T5 token embeddings to Qwen3 hidden states. |
| | """ |
| |
|
| | model_cpu_offload_seq = "text_encoder->llm_adapter->transformer->vae" |
| | _callback_tensor_inputs = ["latents", "prompt_embeds", "negative_prompt_embeds"] |
| |
|
| | def __init__( |
| | self, |
| | text_encoder: PreTrainedModel, |
| | tokenizer: PreTrainedTokenizerFast, |
| | t5_tokenizer: PreTrainedTokenizerFast, |
| | llm_adapter, |
| | transformer: CosmosTransformer3DModel, |
| | vae: AutoencoderKLWan, |
| | scheduler: FlowMatchEulerDiscreteScheduler, |
| | ): |
| | super().__init__() |
| |
|
| | self.register_modules( |
| | text_encoder=text_encoder, |
| | tokenizer=tokenizer, |
| | t5_tokenizer=t5_tokenizer, |
| | llm_adapter=llm_adapter, |
| | transformer=transformer, |
| | vae=vae, |
| | scheduler=scheduler, |
| | ) |
| |
|
| | self.vae_scale_factor_temporal = 2 ** sum(self.vae.temperal_downsample) if getattr(self, "vae", None) else 4 |
| | self.vae_scale_factor_spatial = 2 ** len(self.vae.temperal_downsample) if getattr(self, "vae", None) else 8 |
| | self.video_processor = VideoProcessor(vae_scale_factor=self.vae_scale_factor_spatial) |
| |
|
| | def _encode_prompt( |
| | self, |
| | prompt: Union[str, List[str]], |
| | device: torch.device, |
| | dtype: torch.dtype, |
| | max_sequence_length: int = 512, |
| | ): |
| | """Encode prompt through Qwen3 and run LLM adapter with T5 token IDs.""" |
| | prompt = [prompt] if isinstance(prompt, str) else prompt |
| | batch_size = len(prompt) |
| |
|
| | |
| | all_empty = all(p.strip() == "" for p in prompt) |
| | if all_empty: |
| | return torch.zeros(batch_size, 512, self.llm_adapter.config.target_dim, device=device, dtype=dtype) |
| |
|
| | |
| | qwen_inputs = self.tokenizer( |
| | prompt, |
| | padding=True, |
| | truncation=True, |
| | max_length=max_sequence_length, |
| | return_tensors="pt", |
| | ) |
| | qwen_input_ids = qwen_inputs.input_ids.to(device) |
| | qwen_attention_mask = qwen_inputs.attention_mask.to(device) |
| |
|
| | |
| | qwen_outputs = self.text_encoder( |
| | input_ids=qwen_input_ids, |
| | attention_mask=qwen_attention_mask, |
| | ) |
| | qwen_hidden_states = qwen_outputs.last_hidden_state.to(dtype=dtype) |
| |
|
| | |
| | t5_inputs = self.t5_tokenizer( |
| | prompt, |
| | padding=True, |
| | truncation=True, |
| | max_length=max_sequence_length, |
| | return_tensors="pt", |
| | ) |
| | t5_input_ids = t5_inputs.input_ids.to(device) |
| |
|
| | |
| | adapted_embeds = self.llm_adapter( |
| | source_hidden_states=qwen_hidden_states, |
| | target_input_ids=t5_input_ids, |
| | ) |
| |
|
| | |
| | if adapted_embeds.shape[1] < 512: |
| | adapted_embeds = torch.nn.functional.pad( |
| | adapted_embeds, (0, 0, 0, 512 - adapted_embeds.shape[1]) |
| | ) |
| |
|
| | return adapted_embeds |
| |
|
| | def encode_prompt( |
| | self, |
| | prompt: Union[str, List[str]], |
| | negative_prompt: Optional[Union[str, List[str]]] = None, |
| | do_classifier_free_guidance: bool = True, |
| | num_images_per_prompt: int = 1, |
| | prompt_embeds: Optional[torch.Tensor] = None, |
| | negative_prompt_embeds: Optional[torch.Tensor] = None, |
| | max_sequence_length: int = 512, |
| | device: Optional[torch.device] = None, |
| | dtype: Optional[torch.dtype] = None, |
| | ): |
| | device = device or self._execution_device |
| | dtype = dtype or self.text_encoder.dtype |
| | prompt = [prompt] if isinstance(prompt, str) else prompt |
| |
|
| | if prompt is not None: |
| | batch_size = len(prompt) |
| | else: |
| | batch_size = prompt_embeds.shape[0] |
| |
|
| | if prompt_embeds is None: |
| | prompt_embeds = self._encode_prompt(prompt, device, dtype, max_sequence_length) |
| | _, seq_len, _ = prompt_embeds.shape |
| | prompt_embeds = prompt_embeds.repeat(1, num_images_per_prompt, 1) |
| | prompt_embeds = prompt_embeds.view(batch_size * num_images_per_prompt, seq_len, -1) |
| |
|
| | if do_classifier_free_guidance and negative_prompt_embeds is None: |
| | negative_prompt = negative_prompt or "" |
| | negative_prompt = batch_size * [negative_prompt] if isinstance(negative_prompt, str) else negative_prompt |
| | negative_prompt_embeds = self._encode_prompt(negative_prompt, device, dtype, max_sequence_length) |
| | _, seq_len, _ = negative_prompt_embeds.shape |
| | negative_prompt_embeds = negative_prompt_embeds.repeat(1, num_images_per_prompt, 1) |
| | negative_prompt_embeds = negative_prompt_embeds.view(batch_size * num_images_per_prompt, seq_len, -1) |
| |
|
| | return prompt_embeds, negative_prompt_embeds |
| |
|
| | def prepare_latents( |
| | self, |
| | batch_size: int, |
| | num_channels_latents: int, |
| | height: int, |
| | width: int, |
| | num_frames: int = 1, |
| | dtype: torch.dtype = None, |
| | device: torch.device = None, |
| | generator=None, |
| | latents: torch.Tensor = None, |
| | ): |
| | num_latent_frames = (num_frames - 1) // self.vae_scale_factor_temporal + 1 |
| | latent_height = height // self.vae_scale_factor_spatial |
| | latent_width = width // self.vae_scale_factor_spatial |
| |
|
| | if latents is not None: |
| | return latents.to(device=device, dtype=dtype) |
| |
|
| | shape = (batch_size, num_channels_latents, num_latent_frames, latent_height, latent_width) |
| | latents = randn_tensor(shape, generator=generator, device=device, dtype=dtype) |
| | return latents |
| |
|
| | def check_inputs(self, prompt, height, width, prompt_embeds=None): |
| | if height % 16 != 0 or width % 16 != 0: |
| | raise ValueError(f"`height` and `width` have to be divisible by 16 but are {height} and {width}.") |
| | if prompt is not None and prompt_embeds is not None: |
| | raise ValueError("Cannot forward both `prompt` and `prompt_embeds`.") |
| | elif prompt is None and prompt_embeds is None: |
| | raise ValueError("Provide either `prompt` or `prompt_embeds`.") |
| |
|
| | @property |
| | def guidance_scale(self): |
| | return self._guidance_scale |
| |
|
| | @property |
| | def do_classifier_free_guidance(self): |
| | return self._guidance_scale > 1.0 |
| |
|
| | @property |
| | def num_timesteps(self): |
| | return self._num_timesteps |
| |
|
| | @property |
| | def interrupt(self): |
| | return self._interrupt |
| |
|
| | @torch.no_grad() |
| | def __call__( |
| | self, |
| | prompt: Union[str, List[str]] = None, |
| | negative_prompt: Optional[Union[str, List[str]]] = None, |
| | height: int = 768, |
| | width: int = 1360, |
| | num_inference_steps: int = 35, |
| | guidance_scale: float = 7.0, |
| | num_images_per_prompt: Optional[int] = 1, |
| | generator: Optional[Union[torch.Generator, List[torch.Generator]]] = None, |
| | latents: Optional[torch.Tensor] = None, |
| | prompt_embeds: Optional[torch.Tensor] = None, |
| | negative_prompt_embeds: Optional[torch.Tensor] = None, |
| | output_type: Optional[str] = "pil", |
| | return_dict: bool = True, |
| | callback_on_step_end: Optional[ |
| | Union[Callable[[int, int, Dict], None], PipelineCallback, MultiPipelineCallbacks] |
| | ] = None, |
| | callback_on_step_end_tensor_inputs: List[str] = ["latents"], |
| | max_sequence_length: int = 512, |
| | ): |
| | if isinstance(callback_on_step_end, (PipelineCallback, MultiPipelineCallbacks)): |
| | callback_on_step_end_tensor_inputs = callback_on_step_end.tensor_inputs |
| |
|
| | num_frames = 1 |
| |
|
| | self.check_inputs(prompt, height, width, prompt_embeds) |
| | self._guidance_scale = guidance_scale |
| | self._current_timestep = None |
| | self._interrupt = False |
| |
|
| | device = self._execution_device |
| |
|
| | if prompt is not None and isinstance(prompt, str): |
| | batch_size = 1 |
| | elif prompt is not None and isinstance(prompt, list): |
| | batch_size = len(prompt) |
| | else: |
| | batch_size = prompt_embeds.shape[0] |
| |
|
| | |
| | prompt_embeds, negative_prompt_embeds = self.encode_prompt( |
| | prompt=prompt, |
| | negative_prompt=negative_prompt, |
| | do_classifier_free_guidance=self.do_classifier_free_guidance, |
| | num_images_per_prompt=num_images_per_prompt, |
| | prompt_embeds=prompt_embeds, |
| | negative_prompt_embeds=negative_prompt_embeds, |
| | device=device, |
| | max_sequence_length=max_sequence_length, |
| | ) |
| |
|
| | |
| | timesteps, num_inference_steps = retrieve_timesteps( |
| | self.scheduler, num_inference_steps=num_inference_steps, device=device |
| | ) |
| |
|
| | |
| | transformer_dtype = self.transformer.dtype |
| | num_channels_latents = self.transformer.config.in_channels |
| | latents = self.prepare_latents( |
| | batch_size * num_images_per_prompt, |
| | num_channels_latents, |
| | height, |
| | width, |
| | num_frames, |
| | torch.float32, |
| | device, |
| | generator, |
| | latents, |
| | ) |
| |
|
| | padding_mask = latents.new_zeros(1, 1, height, width, dtype=transformer_dtype) |
| |
|
| | |
| | |
| | |
| | |
| | |
| | num_warmup_steps = len(timesteps) - num_inference_steps * self.scheduler.order |
| | self._num_timesteps = len(timesteps) |
| |
|
| | with self.progress_bar(total=num_inference_steps) as progress_bar: |
| | for i, t in enumerate(timesteps): |
| | if self.interrupt: |
| | continue |
| |
|
| | self._current_timestep = t |
| | sigma = self.scheduler.sigmas[i] |
| |
|
| | |
| | timestep = sigma.expand(latents.shape[0]).to(transformer_dtype) |
| | latent_model_input = latents.to(transformer_dtype) |
| |
|
| | |
| | velocity = self.transformer( |
| | hidden_states=latent_model_input, |
| | timestep=timestep, |
| | encoder_hidden_states=prompt_embeds, |
| | padding_mask=padding_mask, |
| | return_dict=False, |
| | )[0].float() |
| |
|
| | if self.do_classifier_free_guidance: |
| | velocity_uncond = self.transformer( |
| | hidden_states=latent_model_input, |
| | timestep=timestep, |
| | encoder_hidden_states=negative_prompt_embeds, |
| | padding_mask=padding_mask, |
| | return_dict=False, |
| | )[0].float() |
| | velocity = velocity_uncond + self.guidance_scale * (velocity - velocity_uncond) |
| |
|
| | |
| | latents = self.scheduler.step(velocity, t, latents, return_dict=False)[0] |
| |
|
| | if callback_on_step_end is not None: |
| | callback_kwargs = {} |
| | for k in callback_on_step_end_tensor_inputs: |
| | callback_kwargs[k] = locals()[k] |
| | callback_outputs = callback_on_step_end(self, i, t, callback_kwargs) |
| | latents = callback_outputs.pop("latents", latents) |
| | prompt_embeds = callback_outputs.pop("prompt_embeds", prompt_embeds) |
| | negative_prompt_embeds = callback_outputs.pop("negative_prompt_embeds", negative_prompt_embeds) |
| |
|
| | if i == len(timesteps) - 1 or ((i + 1) > num_warmup_steps and (i + 1) % self.scheduler.order == 0): |
| | progress_bar.update() |
| |
|
| | self._current_timestep = None |
| |
|
| | if not output_type == "latent": |
| | latents_mean = ( |
| | torch.tensor(self.vae.config.latents_mean) |
| | .view(1, self.vae.config.z_dim, 1, 1, 1) |
| | .to(latents.device, latents.dtype) |
| | ) |
| | latents_std = 1.0 / torch.tensor(self.vae.config.latents_std).view(1, self.vae.config.z_dim, 1, 1, 1).to( |
| | latents.device, latents.dtype |
| | ) |
| | latents = latents / latents_std + latents_mean |
| | video = self.vae.decode(latents.to(self.vae.dtype), return_dict=False)[0] |
| | video = self.video_processor.postprocess_video(video, output_type=output_type) |
| | image = [batch[0] for batch in video] |
| | if isinstance(video, torch.Tensor): |
| | image = torch.stack(image) |
| | elif isinstance(video, np.ndarray): |
| | image = np.stack(image) |
| | else: |
| | image = latents[:, :, 0] |
| |
|
| | self.maybe_free_model_hooks() |
| |
|
| | if not return_dict: |
| | return (image,) |
| |
|
| | return CosmosImagePipelineOutput(images=image) |
| |
|