| import torch, math |
| from PIL import Image |
| from typing import Union |
| from tqdm import tqdm |
| from einops import rearrange |
| import numpy as np |
| from typing import Union, List, Optional, Tuple |
|
|
| from ..diffusion import FlowMatchScheduler |
| from ..core import ModelConfig, gradient_checkpoint_forward |
| from ..diffusion.base_pipeline import BasePipeline, PipelineUnit, ControlNetInput |
|
|
| from transformers import AutoProcessor |
| from ..models.flux2_text_encoder import Flux2TextEncoder |
| from ..models.flux2_dit import Flux2DiT |
| from ..models.flux2_vae import Flux2VAE |
|
|
|
|
| class Flux2ImagePipeline(BasePipeline): |
|
|
| def __init__(self, device="cuda", torch_dtype=torch.bfloat16): |
| super().__init__( |
| device=device, torch_dtype=torch_dtype, |
| height_division_factor=16, width_division_factor=16, |
| ) |
| self.scheduler = FlowMatchScheduler("FLUX.2") |
| self.text_encoder: Flux2TextEncoder = None |
| self.dit: Flux2DiT = None |
| self.vae: Flux2VAE = None |
| self.tokenizer: AutoProcessor = None |
| self.in_iteration_models = ("dit",) |
| self.units = [ |
| Flux2Unit_ShapeChecker(), |
| Flux2Unit_PromptEmbedder(), |
| Flux2Unit_NoiseInitializer(), |
| Flux2Unit_InputImageEmbedder(), |
| Flux2Unit_ImageIDs(), |
| ] |
| self.model_fn = model_fn_flux2 |
| |
| |
| @staticmethod |
| def from_pretrained( |
| torch_dtype: torch.dtype = torch.bfloat16, |
| device: Union[str, torch.device] = "cuda", |
| model_configs: list[ModelConfig] = [], |
| tokenizer_config: ModelConfig = ModelConfig(model_id="black-forest-labs/FLUX.2-dev", origin_file_pattern="tokenizer/"), |
| vram_limit: float = None, |
| ): |
| |
| pipe = Flux2ImagePipeline(device=device, torch_dtype=torch_dtype) |
| model_pool = pipe.download_and_load_models(model_configs, vram_limit) |
| |
| |
| pipe.text_encoder = model_pool.fetch_model("flux2_text_encoder") |
| pipe.dit = model_pool.fetch_model("flux2_dit") |
| pipe.vae = model_pool.fetch_model("flux2_vae") |
| if tokenizer_config is not None: |
| tokenizer_config.download_if_necessary() |
| pipe.tokenizer = AutoProcessor.from_pretrained(tokenizer_config.path) |
| |
| |
| pipe.vram_management_enabled = pipe.check_vram_management_state() |
| return pipe |
| |
| |
| @torch.no_grad() |
| def __call__( |
| self, |
| |
| prompt: str, |
| negative_prompt: str = "", |
| cfg_scale: float = 1.0, |
| embedded_guidance: float = 4.0, |
| |
| input_image: Image.Image = None, |
| denoising_strength: float = 1.0, |
| |
| height: int = 1024, |
| width: int = 1024, |
| |
| seed: int = None, |
| rand_device: str = "cpu", |
| |
| num_inference_steps: int = 30, |
| |
| progress_bar_cmd = tqdm, |
| ): |
| self.scheduler.set_timesteps(num_inference_steps, denoising_strength=denoising_strength, dynamic_shift_len=height//16*width//16) |
|
|
| |
| inputs_posi = { |
| "prompt": prompt, |
| } |
| inputs_nega = { |
| "negative_prompt": negative_prompt, |
| } |
| inputs_shared = { |
| "cfg_scale": cfg_scale, "embedded_guidance": embedded_guidance, |
| "input_image": input_image, "denoising_strength": denoising_strength, |
| "height": height, "width": width, |
| "seed": seed, "rand_device": rand_device, |
| "num_inference_steps": num_inference_steps, |
| } |
| for unit in self.units: |
| inputs_shared, inputs_posi, inputs_nega = self.unit_runner(unit, self, inputs_shared, inputs_posi, inputs_nega) |
|
|
| |
| self.load_models_to_device(self.in_iteration_models) |
| models = {name: getattr(self, name) for name in self.in_iteration_models} |
| for progress_id, timestep in enumerate(progress_bar_cmd(self.scheduler.timesteps)): |
| timestep = timestep.unsqueeze(0).to(dtype=self.torch_dtype, device=self.device) |
| noise_pred = self.cfg_guided_model_fn( |
| self.model_fn, cfg_scale, |
| inputs_shared, inputs_posi, inputs_nega, |
| **models, timestep=timestep, progress_id=progress_id |
| ) |
| inputs_shared["latents"] = self.step(self.scheduler, progress_id=progress_id, noise_pred=noise_pred, **inputs_shared) |
| |
| |
| self.load_models_to_device(['vae']) |
| latents = rearrange(inputs_shared["latents"], "B (H W) C -> B C H W", H=inputs_shared["height"]//16, W=inputs_shared["width"]//16) |
| image = self.vae.decode(latents) |
| image = self.vae_output_to_image(image) |
| self.load_models_to_device([]) |
|
|
| return image |
|
|
|
|
| class Flux2Unit_ShapeChecker(PipelineUnit): |
| def __init__(self): |
| super().__init__( |
| input_params=("height", "width"), |
| output_params=("height", "width"), |
| ) |
|
|
| def process(self, pipe: Flux2ImagePipeline, height, width): |
| height, width = pipe.check_resize_height_width(height, width) |
| return {"height": height, "width": width} |
|
|
|
|
| class Flux2Unit_PromptEmbedder(PipelineUnit): |
| def __init__(self): |
| super().__init__( |
| seperate_cfg=True, |
| input_params_posi={"prompt": "prompt"}, |
| input_params_nega={"prompt": "negative_prompt"}, |
| output_params=("prompt_emb", "prompt_emb_mask"), |
| onload_model_names=("text_encoder",) |
| ) |
| self.system_message = "You are an AI that reasons about image descriptions. You give structured responses focusing on object relationships, object attribution and actions without speculation." |
|
|
| def format_text_input(self, prompts: List[str], system_message: str = None): |
| |
| |
| |
| cleaned_txt = [prompt.replace("[IMG]", "") for prompt in prompts] |
|
|
| return [ |
| [ |
| { |
| "role": "system", |
| "content": [{"type": "text", "text": system_message}], |
| }, |
| {"role": "user", "content": [{"type": "text", "text": prompt}]}, |
| ] |
| for prompt in cleaned_txt |
| ] |
|
|
| def get_mistral_3_small_prompt_embeds( |
| self, |
| text_encoder, |
| tokenizer, |
| prompt: Union[str, List[str]], |
| dtype: Optional[torch.dtype] = None, |
| device: Optional[torch.device] = None, |
| max_sequence_length: int = 512, |
| |
| system_message: str = "You are an AI that reasons about image descriptions. You give structured responses focusing on object relationships, object attribution and actions without speculation.", |
| |
| hidden_states_layers: List[int] = (10, 20, 30), |
| ): |
| dtype = text_encoder.dtype if dtype is None else dtype |
| device = text_encoder.device if device is None else device |
|
|
| prompt = [prompt] if isinstance(prompt, str) else prompt |
|
|
| |
| messages_batch = self.format_text_input(prompts=prompt, system_message=system_message) |
|
|
| |
| inputs = tokenizer.apply_chat_template( |
| messages_batch, |
| add_generation_prompt=False, |
| tokenize=True, |
| return_dict=True, |
| return_tensors="pt", |
| padding="max_length", |
| truncation=True, |
| max_length=max_sequence_length, |
| ) |
|
|
| |
| input_ids = inputs["input_ids"].to(device) |
| attention_mask = inputs["attention_mask"].to(device) |
|
|
| |
| output = text_encoder( |
| input_ids=input_ids, |
| attention_mask=attention_mask, |
| output_hidden_states=True, |
| use_cache=False, |
| ) |
|
|
| |
| out = torch.stack([output.hidden_states[k] for k in hidden_states_layers], dim=1) |
| out = out.to(dtype=dtype, device=device) |
|
|
| batch_size, num_channels, seq_len, hidden_dim = out.shape |
| prompt_embeds = out.permute(0, 2, 1, 3).reshape(batch_size, seq_len, num_channels * hidden_dim) |
|
|
| return prompt_embeds |
| |
| def prepare_text_ids( |
| self, |
| x: torch.Tensor, |
| t_coord: Optional[torch.Tensor] = None, |
| ): |
| B, L, _ = x.shape |
| out_ids = [] |
|
|
| for i in range(B): |
| t = torch.arange(1) if t_coord is None else t_coord[i] |
| h = torch.arange(1) |
| w = torch.arange(1) |
| l = torch.arange(L) |
|
|
| coords = torch.cartesian_prod(t, h, w, l) |
| out_ids.append(coords) |
|
|
| return torch.stack(out_ids) |
|
|
| def encode_prompt( |
| self, |
| text_encoder, |
| tokenizer, |
| prompt: Union[str, List[str]], |
| dtype = None, |
| device: Optional[torch.device] = None, |
| num_images_per_prompt: int = 1, |
| prompt_embeds: Optional[torch.Tensor] = None, |
| max_sequence_length: int = 512, |
| text_encoder_out_layers: Tuple[int] = (10, 20, 30), |
| ): |
| prompt = [prompt] if isinstance(prompt, str) else prompt |
|
|
| if prompt_embeds is None: |
| prompt_embeds = self.get_mistral_3_small_prompt_embeds( |
| text_encoder=text_encoder, |
| tokenizer=tokenizer, |
| prompt=prompt, |
| dtype=dtype, |
| device=device, |
| max_sequence_length=max_sequence_length, |
| system_message=self.system_message, |
| hidden_states_layers=text_encoder_out_layers, |
| ) |
|
|
| batch_size, seq_len, _ = prompt_embeds.shape |
| prompt_embeds = prompt_embeds.repeat(1, num_images_per_prompt, 1) |
| prompt_embeds = prompt_embeds.view(batch_size * num_images_per_prompt, seq_len, -1) |
|
|
| text_ids = self.prepare_text_ids(prompt_embeds) |
| text_ids = text_ids.to(device) |
| return prompt_embeds, text_ids |
|
|
| def process(self, pipe: Flux2ImagePipeline, prompt): |
| pipe.load_models_to_device(self.onload_model_names) |
| prompt_embeds, text_ids = self.encode_prompt( |
| pipe.text_encoder, pipe.tokenizer, prompt, |
| dtype=pipe.torch_dtype, device=pipe.device, |
| ) |
| return {"prompt_embeds": prompt_embeds, "text_ids": text_ids} |
|
|
|
|
| class Flux2Unit_NoiseInitializer(PipelineUnit): |
| def __init__(self): |
| super().__init__( |
| input_params=("height", "width", "seed", "rand_device"), |
| output_params=("noise",), |
| ) |
|
|
| def process(self, pipe: Flux2ImagePipeline, height, width, seed, rand_device): |
| noise = pipe.generate_noise((1, 128, height//16, width//16), seed=seed, rand_device=rand_device, rand_torch_dtype=pipe.torch_dtype) |
| noise = noise.reshape(1, 128, height//16 * width//16).permute(0, 2, 1) |
| return {"noise": noise} |
|
|
|
|
| class Flux2Unit_InputImageEmbedder(PipelineUnit): |
| def __init__(self): |
| super().__init__( |
| input_params=("input_image", "noise"), |
| output_params=("latents", "input_latents"), |
| onload_model_names=("vae",) |
| ) |
|
|
| def process(self, pipe: Flux2ImagePipeline, input_image, noise): |
| if input_image is None: |
| return {"latents": noise, "input_latents": None} |
| pipe.load_models_to_device(['vae']) |
| image = pipe.preprocess_image(input_image) |
| input_latents = pipe.vae.encode(image) |
| input_latents = rearrange(input_latents, "B C H W -> B (H W) C") |
| if pipe.scheduler.training: |
| return {"latents": noise, "input_latents": input_latents} |
| else: |
| latents = pipe.scheduler.add_noise(input_latents, noise, timestep=pipe.scheduler.timesteps[0]) |
| return {"latents": latents, "input_latents": input_latents} |
|
|
|
|
| class Flux2Unit_ImageIDs(PipelineUnit): |
| def __init__(self): |
| super().__init__( |
| input_params=("height", "width"), |
| output_params=("image_ids",), |
| ) |
|
|
| def prepare_latent_ids(self, height, width): |
| t = torch.arange(1) |
| h = torch.arange(height) |
| w = torch.arange(width) |
| l = torch.arange(1) |
|
|
| |
| latent_ids = torch.cartesian_prod(t, h, w, l) |
|
|
| |
| latent_ids = latent_ids.unsqueeze(0).expand(1, -1, -1) |
|
|
| return latent_ids |
|
|
| def process(self, pipe: Flux2ImagePipeline, height, width): |
| image_ids = self.prepare_latent_ids(height // 16, width // 16).to(pipe.device) |
| return {"image_ids": image_ids} |
|
|
|
|
| def model_fn_flux2( |
| dit: Flux2DiT, |
| latents=None, |
| timestep=None, |
| embedded_guidance=None, |
| prompt_embeds=None, |
| text_ids=None, |
| image_ids=None, |
| use_gradient_checkpointing=False, |
| use_gradient_checkpointing_offload=False, |
| **kwargs, |
| ): |
| embedded_guidance = torch.tensor([embedded_guidance], device=latents.device) |
| model_output = dit( |
| hidden_states=latents, |
| timestep=timestep / 1000, |
| guidance=embedded_guidance, |
| encoder_hidden_states=prompt_embeds, |
| txt_ids=text_ids, |
| img_ids=image_ids, |
| use_gradient_checkpointing=use_gradient_checkpointing, |
| use_gradient_checkpointing_offload=use_gradient_checkpointing_offload, |
| ) |
| return model_output |
|
|