sample_id
stringlengths
21
196
text
stringlengths
105
936k
metadata
dict
category
stringclasses
6 values
Comfy-Org/ComfyUI:comfy_api_nodes/util/common_exceptions.py
class NetworkError(Exception): """Base exception for network-related errors with diagnostic information.""" class LocalNetworkError(NetworkError): """Exception raised when local network connectivity issues are detected.""" class ApiServerError(NetworkError): """Exception raised when the API server is unreachable but internet is working.""" class ProcessingInterrupted(Exception): """Operation was interrupted by user/runtime via processing_interrupted()."""
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api_nodes/util/common_exceptions.py", "license": "GNU General Public License v3.0", "lines": 8, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
documentation
Comfy-Org/ComfyUI:comfy_api_nodes/util/conversions.py
import base64 import logging import math import mimetypes import uuid from io import BytesIO import av import numpy as np import torch from PIL import Image from comfy.utils import common_upscale from comfy_api.latest import Input, InputImpl, Types from ._helpers import mimetype_to_extension def bytesio_to_image_tensor(image_bytesio: BytesIO, mode: str = "RGBA") -> torch.Tensor: """Converts image data from BytesIO to a torch.Tensor. Args: image_bytesio: BytesIO object containing the image data. mode: The PIL mode to convert the image to (e.g., "RGB", "RGBA"). Returns: A torch.Tensor representing the image (1, H, W, C). Raises: PIL.UnidentifiedImageError: If the image data cannot be identified. ValueError: If the specified mode is invalid. """ image = Image.open(image_bytesio) image = image.convert(mode) image_array = np.array(image).astype(np.float32) / 255.0 return torch.from_numpy(image_array).unsqueeze(0) def image_tensor_pair_to_batch(image1: torch.Tensor, image2: torch.Tensor) -> torch.Tensor: """ Converts a pair of image tensors to a batch tensor. If the images are not the same size, the smaller image is resized to match the larger image. """ if image1.shape[1:] != image2.shape[1:]: image2 = common_upscale( image2.movedim(-1, 1), image1.shape[2], image1.shape[1], "bilinear", "center", ).movedim(1, -1) return torch.cat((image1, image2), dim=0) def tensor_to_bytesio( image: torch.Tensor, *, total_pixels: int | None = 2048 * 2048, mime_type: str | None = "image/png", ) -> BytesIO: """Converts a torch.Tensor image to a named BytesIO object. Args: image: Input torch.Tensor image. total_pixels: Maximum total pixels for downscaling. If None, no downscaling is performed. mime_type: Target image MIME type (e.g., 'image/png', 'image/jpeg', 'image/webp', 'video/mp4'). Returns: Named BytesIO object containing the image data, with pointer set to the start of buffer. """ if not mime_type: mime_type = "image/png" pil_image = tensor_to_pil(image, total_pixels=total_pixels) img_binary = pil_to_bytesio(pil_image, mime_type=mime_type) img_binary.name = f"{uuid.uuid4()}.{mimetype_to_extension(mime_type)}" return img_binary def tensor_to_pil(image: torch.Tensor, total_pixels: int | None = 2048 * 2048) -> Image.Image: """Converts a single torch.Tensor image [H, W, C] to a PIL Image, optionally downscaling.""" if len(image.shape) > 3: image = image[0] # TODO: remove alpha if not allowed and present input_tensor = image.cpu() if total_pixels is not None: input_tensor = downscale_image_tensor(input_tensor.unsqueeze(0), total_pixels=total_pixels).squeeze() image_np = (input_tensor.numpy() * 255).astype(np.uint8) img = Image.fromarray(image_np) return img def tensor_to_base64_string( image_tensor: torch.Tensor, total_pixels: int | None = 2048 * 2048, mime_type: str = "image/png", ) -> str: """Convert [B, H, W, C] or [H, W, C] tensor to a base64 string. Args: image_tensor: Input torch.Tensor image. total_pixels: Maximum total pixels for downscaling. If None, no downscaling is performed. mime_type: Target image MIME type (e.g., 'image/png', 'image/jpeg', 'image/webp', 'video/mp4'). Returns: Base64 encoded string of the image. """ pil_image = tensor_to_pil(image_tensor, total_pixels=total_pixels) img_byte_arr = pil_to_bytesio(pil_image, mime_type=mime_type) img_bytes = img_byte_arr.getvalue() # Encode bytes to base64 string base64_encoded_string = base64.b64encode(img_bytes).decode("utf-8") return base64_encoded_string def pil_to_bytesio(img: Image.Image, mime_type: str = "image/png") -> BytesIO: """Converts a PIL Image to a BytesIO object.""" if not mime_type: mime_type = "image/png" img_byte_arr = BytesIO() # Derive PIL format from MIME type (e.g., 'image/png' -> 'PNG') pil_format = mime_type.split("/")[-1].upper() if pil_format == "JPG": pil_format = "JPEG" img.save(img_byte_arr, format=pil_format) img_byte_arr.seek(0) return img_byte_arr def downscale_image_tensor(image: torch.Tensor, total_pixels: int = 1536 * 1024) -> torch.Tensor: """Downscale input image tensor to roughly the specified total pixels.""" samples = image.movedim(-1, 1) total = int(total_pixels) scale_by = math.sqrt(total / (samples.shape[3] * samples.shape[2])) if scale_by >= 1: return image width = round(samples.shape[3] * scale_by) height = round(samples.shape[2] * scale_by) s = common_upscale(samples, width, height, "lanczos", "disabled") s = s.movedim(1, -1) return s def downscale_image_tensor_by_max_side(image: torch.Tensor, *, max_side: int) -> torch.Tensor: """Downscale input image tensor so the largest dimension is at most max_side pixels.""" samples = image.movedim(-1, 1) height, width = samples.shape[2], samples.shape[3] max_dim = max(width, height) if max_dim <= max_side: return image scale_by = max_side / max_dim new_width = round(width * scale_by) new_height = round(height * scale_by) s = common_upscale(samples, new_width, new_height, "lanczos", "disabled") s = s.movedim(1, -1) return s def tensor_to_data_uri( image_tensor: torch.Tensor, total_pixels: int | None = 2048 * 2048, mime_type: str = "image/png", ) -> str: """Converts a tensor image to a Data URI string. Args: image_tensor: Input torch.Tensor image. total_pixels: Maximum total pixels for downscaling. If None, no downscaling is performed. mime_type: Target image MIME type (e.g., 'image/png', 'image/jpeg', 'image/webp'). Returns: Data URI string (e.g., 'data:image/png;base64,...'). """ base64_string = tensor_to_base64_string(image_tensor, total_pixels, mime_type) return f"data:{mime_type};base64,{base64_string}" def audio_to_base64_string(audio: Input.Audio, container_format: str = "mp4", codec_name: str = "aac") -> str: """Converts an audio input to a base64 string.""" sample_rate: int = audio["sample_rate"] waveform: torch.Tensor = audio["waveform"] audio_data_np = audio_tensor_to_contiguous_ndarray(waveform) audio_bytes_io = audio_ndarray_to_bytesio(audio_data_np, sample_rate, container_format, codec_name) audio_bytes = audio_bytes_io.getvalue() return base64.b64encode(audio_bytes).decode("utf-8") def video_to_base64_string( video: Input.Video, container_format: Types.VideoContainer | None = None, codec: Types.VideoCodec | None = None, ) -> str: """ Converts a video input to a base64 string. Args: video: The video input to convert container_format: Optional container format to use (defaults to video.container if available) codec: Optional codec to use (defaults to video.codec if available) """ video_bytes_io = BytesIO() video.save_to( video_bytes_io, format=container_format or getattr(video, "container", Types.VideoContainer.MP4), codec=codec or getattr(video, "codec", Types.VideoCodec.H264), ) video_bytes_io.seek(0) return base64.b64encode(video_bytes_io.getvalue()).decode("utf-8") def audio_ndarray_to_bytesio( audio_data_np: np.ndarray, sample_rate: int, container_format: str = "mp4", codec_name: str = "aac", ) -> BytesIO: """ Encodes a numpy array of audio data into a BytesIO object. """ audio_bytes_io = BytesIO() with av.open(audio_bytes_io, mode="w", format=container_format) as output_container: audio_stream = output_container.add_stream(codec_name, rate=sample_rate) frame = av.AudioFrame.from_ndarray( audio_data_np, format="fltp", layout="stereo" if audio_data_np.shape[0] > 1 else "mono", ) frame.sample_rate = sample_rate frame.pts = 0 for packet in audio_stream.encode(frame): output_container.mux(packet) # Flush stream for packet in audio_stream.encode(None): output_container.mux(packet) audio_bytes_io.seek(0) return audio_bytes_io def audio_tensor_to_contiguous_ndarray(waveform: torch.Tensor) -> np.ndarray: """ Prepares audio waveform for av library by converting to a contiguous numpy array. Args: waveform: a tensor of shape (1, channels, samples) derived from a Comfy `AUDIO` type. Returns: Contiguous numpy array of the audio waveform. If the audio was batched, the first item is taken. """ if waveform.ndim != 3 or waveform.shape[0] != 1: raise ValueError("Expected waveform tensor shape (1, channels, samples)") # If batch is > 1, take first item if waveform.shape[0] > 1: waveform = waveform[0] # Prepare for av: remove batch dim, move to CPU, make contiguous, convert to numpy array audio_data_np = waveform.squeeze(0).cpu().contiguous().numpy() if audio_data_np.dtype != np.float32: audio_data_np = audio_data_np.astype(np.float32) return audio_data_np def audio_input_to_mp3(audio: Input.Audio) -> BytesIO: waveform = audio["waveform"].cpu() output_buffer = BytesIO() output_container = av.open(output_buffer, mode="w", format="mp3") out_stream = output_container.add_stream("libmp3lame", rate=audio["sample_rate"]) out_stream.bit_rate = 320000 frame = av.AudioFrame.from_ndarray( waveform.movedim(0, 1).reshape(1, -1).float().numpy(), format="flt", layout="mono" if waveform.shape[0] == 1 else "stereo", ) frame.sample_rate = audio["sample_rate"] frame.pts = 0 output_container.mux(out_stream.encode(frame)) output_container.mux(out_stream.encode(None)) output_container.close() output_buffer.seek(0) return output_buffer def trim_video(video: Input.Video, duration_sec: float) -> Input.Video: """ Returns a new VideoInput object trimmed from the beginning to the specified duration, using av to avoid loading entire video into memory. Args: video: Input video to trim duration_sec: Duration in seconds to keep from the beginning Returns: VideoFromFile object that owns the output buffer """ output_buffer = BytesIO() input_container = None output_container = None try: # Get the stream source - this avoids loading entire video into memory # when the source is already a file path input_source = video.get_stream_source() # Open containers input_container = av.open(input_source, mode="r") output_container = av.open(output_buffer, mode="w", format="mp4") # Set up output streams for re-encoding video_stream = None audio_stream = None for stream in input_container.streams: logging.info("Found stream: type=%s, class=%s", stream.type, type(stream)) if isinstance(stream, av.VideoStream): # Create output video stream with same parameters video_stream = output_container.add_stream("h264", rate=stream.average_rate) video_stream.width = stream.width video_stream.height = stream.height video_stream.pix_fmt = "yuv420p" logging.info("Added video stream: %sx%s @ %sfps", stream.width, stream.height, stream.average_rate) elif isinstance(stream, av.AudioStream): # Create output audio stream with same parameters audio_stream = output_container.add_stream("aac", rate=stream.sample_rate) audio_stream.sample_rate = stream.sample_rate audio_stream.layout = stream.layout logging.info("Added audio stream: %sHz, %s channels", stream.sample_rate, stream.channels) # Calculate target frame count that's divisible by 16 fps = input_container.streams.video[0].average_rate estimated_frames = int(duration_sec * fps) target_frames = (estimated_frames // 16) * 16 # Round down to nearest multiple of 16 if target_frames == 0: raise ValueError("Video too short: need at least 16 frames for Moonvalley") frame_count = 0 audio_frame_count = 0 # Decode and re-encode video frames if video_stream: for frame in input_container.decode(video=0): if frame_count >= target_frames: break # Re-encode frame for packet in video_stream.encode(frame): output_container.mux(packet) frame_count += 1 # Flush encoder for packet in video_stream.encode(): output_container.mux(packet) logging.info("Encoded %s video frames (target: %s)", frame_count, target_frames) # Decode and re-encode audio frames if audio_stream: input_container.seek(0) # Reset to beginning for audio for frame in input_container.decode(audio=0): if frame.time >= duration_sec: break # Re-encode frame for packet in audio_stream.encode(frame): output_container.mux(packet) audio_frame_count += 1 # Flush encoder for packet in audio_stream.encode(): output_container.mux(packet) logging.info("Encoded %s audio frames", audio_frame_count) # Close containers output_container.close() input_container.close() # Return as VideoFromFile using the buffer output_buffer.seek(0) return InputImpl.VideoFromFile(output_buffer) except Exception as e: # Clean up on error if input_container is not None: input_container.close() if output_container is not None: output_container.close() raise RuntimeError(f"Failed to trim video: {str(e)}") from e def _f32_pcm(wav: torch.Tensor) -> torch.Tensor: """Convert audio to float 32 bits PCM format. Copy-paste from nodes_audio.py file.""" if wav.dtype.is_floating_point: return wav elif wav.dtype == torch.int16: return wav.float() / (2**15) elif wav.dtype == torch.int32: return wav.float() / (2**31) raise ValueError(f"Unsupported wav dtype: {wav.dtype}") def audio_bytes_to_audio_input(audio_bytes: bytes) -> dict: """ Decode any common audio container from bytes using PyAV and return a Comfy AUDIO dict: {"waveform": [1, C, T] float32, "sample_rate": int}. """ with av.open(BytesIO(audio_bytes)) as af: if not af.streams.audio: raise ValueError("No audio stream found in response.") stream = af.streams.audio[0] in_sr = int(stream.codec_context.sample_rate) out_sr = in_sr frames: list[torch.Tensor] = [] n_channels = stream.channels or 1 for frame in af.decode(streams=stream.index): arr = frame.to_ndarray() # shape can be [C, T] or [T, C] or [T] buf = torch.from_numpy(arr) if buf.ndim == 1: buf = buf.unsqueeze(0) # [T] -> [1, T] elif buf.shape[0] != n_channels and buf.shape[-1] == n_channels: buf = buf.transpose(0, 1).contiguous() # [T, C] -> [C, T] elif buf.shape[0] != n_channels: buf = buf.reshape(-1, n_channels).t().contiguous() # fallback to [C, T] frames.append(buf) if not frames: raise ValueError("Decoded zero audio frames.") wav = torch.cat(frames, dim=1) # [C, T] wav = _f32_pcm(wav) return {"waveform": wav.unsqueeze(0).contiguous(), "sample_rate": out_sr} def resize_mask_to_image( mask: torch.Tensor, image: torch.Tensor, upscale_method="nearest-exact", crop="disabled", allow_gradient=True, add_channel_dim=False, ): """Resize mask to be the same dimensions as an image, while maintaining proper format for API calls.""" _, height, width, _ = image.shape mask = mask.unsqueeze(-1) mask = mask.movedim(-1, 1) mask = common_upscale(mask, width=width, height=height, upscale_method=upscale_method, crop=crop) mask = mask.movedim(1, -1) if not add_channel_dim: mask = mask.squeeze(-1) if not allow_gradient: mask = (mask > 0.5).float() return mask def convert_mask_to_image(mask: Input.Image) -> torch.Tensor: """Make mask have the expected amount of dims (4) and channels (3) to be recognized as an image.""" mask = mask.unsqueeze(-1) return torch.cat([mask] * 3, dim=-1) def text_filepath_to_base64_string(filepath: str) -> str: """Converts a text file to a base64 string.""" with open(filepath, "rb") as f: file_content = f.read() return base64.b64encode(file_content).decode("utf-8") def text_filepath_to_data_uri(filepath: str) -> str: """Converts a text file to a data URI.""" base64_string = text_filepath_to_base64_string(filepath) mime_type, _ = mimetypes.guess_type(filepath) if mime_type is None: mime_type = "application/octet-stream" return f"data:{mime_type};base64,{base64_string}"
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api_nodes/util/conversions.py", "license": "GNU General Public License v3.0", "lines": 394, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_api_nodes/util/download_helpers.py
import asyncio import contextlib import uuid from io import BytesIO from pathlib import Path from typing import IO from urllib.parse import urljoin, urlparse import aiohttp import torch from aiohttp.client_exceptions import ClientError, ContentTypeError from comfy_api.latest import IO as COMFY_IO from comfy_api.latest import InputImpl, Types from folder_paths import get_output_directory from . import request_logger from ._helpers import ( default_base_url, get_auth_header, is_processing_interrupted, sleep_with_interrupt, to_aiohttp_url, ) from .client import _diagnose_connectivity from .common_exceptions import ApiServerError, LocalNetworkError, ProcessingInterrupted from .conversions import bytesio_to_image_tensor _RETRY_STATUS = {408, 429, 500, 502, 503, 504} async def download_url_to_bytesio( url: str, dest: BytesIO | IO[bytes] | str | Path | None, *, timeout: float | None = None, max_retries: int = 5, retry_delay: float = 1.0, retry_backoff: float = 2.0, cls: type[COMFY_IO.ComfyNode] = None, ) -> None: """Stream-download a URL to `dest`. `dest` must be one of: - a BytesIO (rewound to 0 after write), - a file-like object opened in binary write mode (must implement .write()), - a filesystem path (str | pathlib.Path), which will be opened with 'wb'. If `url` starts with `/proxy/`, `cls` must be provided so the URL can be expanded to an absolute URL and authentication headers can be applied. Raises: ProcessingInterrupted, LocalNetworkError, ApiServerError, Exception (HTTP and other errors) """ if not isinstance(dest, (str, Path)) and not hasattr(dest, "write"): raise ValueError("dest must be a path (str|Path) or a binary-writable object providing .write().") attempt = 0 delay = retry_delay headers: dict[str, str] = {} parsed_url = urlparse(url) if not parsed_url.scheme and not parsed_url.netloc: # is URL relative? if cls is None: raise ValueError("For relative 'cloud' paths, the `cls` parameter is required.") url = urljoin(default_base_url().rstrip("/") + "/", url.lstrip("/")) headers = get_auth_header(cls) while True: attempt += 1 op_id = _generate_operation_id("GET", url, attempt) timeout_cfg = aiohttp.ClientTimeout(total=timeout) is_path_sink = isinstance(dest, (str, Path)) fhandle = None session: aiohttp.ClientSession | None = None stop_evt: asyncio.Event | None = None monitor_task: asyncio.Task | None = None req_task: asyncio.Task | None = None try: with contextlib.suppress(Exception): request_logger.log_request_response(operation_id=op_id, request_method="GET", request_url=url) session = aiohttp.ClientSession(timeout=timeout_cfg) stop_evt = asyncio.Event() async def _monitor(): try: while not stop_evt.is_set(): if is_processing_interrupted(): return await asyncio.sleep(1.0) except asyncio.CancelledError: return monitor_task = asyncio.create_task(_monitor()) req_task = asyncio.create_task(session.get(to_aiohttp_url(url), headers=headers)) done, pending = await asyncio.wait({req_task, monitor_task}, return_when=asyncio.FIRST_COMPLETED) if monitor_task in done and req_task in pending: req_task.cancel() with contextlib.suppress(Exception): await req_task raise ProcessingInterrupted("Task cancelled") try: resp = await req_task except asyncio.CancelledError: raise ProcessingInterrupted("Task cancelled") from None async with resp: if resp.status >= 400: with contextlib.suppress(Exception): try: body = await resp.json() except (ContentTypeError, ValueError): text = await resp.text() body = text if len(text) <= 4096 else f"[text {len(text)} bytes]" request_logger.log_request_response( operation_id=op_id, request_method="GET", request_url=url, response_status_code=resp.status, response_headers=dict(resp.headers), response_content=body, error_message=f"HTTP {resp.status}", ) if resp.status in _RETRY_STATUS and attempt <= max_retries: await sleep_with_interrupt(delay, cls, None, None, None) delay *= retry_backoff continue raise Exception(f"Failed to download (HTTP {resp.status}).") if is_path_sink: p = Path(str(dest)) with contextlib.suppress(Exception): p.parent.mkdir(parents=True, exist_ok=True) fhandle = open(p, "wb") sink = fhandle else: sink = dest # BytesIO or file-like written = 0 while True: try: chunk = await asyncio.wait_for(resp.content.read(1024 * 1024), timeout=1.0) except asyncio.TimeoutError: chunk = b"" except asyncio.CancelledError: raise ProcessingInterrupted("Task cancelled") from None if is_processing_interrupted(): raise ProcessingInterrupted("Task cancelled") if not chunk: if resp.content.at_eof(): break continue sink.write(chunk) written += len(chunk) if isinstance(dest, BytesIO): with contextlib.suppress(Exception): dest.seek(0) request_logger.log_request_response( operation_id=op_id, request_method="GET", request_url=url, response_status_code=resp.status, response_headers=dict(resp.headers), response_content=f"[streamed {written} bytes to dest]", ) return except asyncio.CancelledError: raise ProcessingInterrupted("Task cancelled") from None except (ClientError, OSError) as e: if attempt <= max_retries: request_logger.log_request_response( operation_id=op_id, request_method="GET", request_url=url, error_message=f"{type(e).__name__}: {str(e)} (will retry)", ) await sleep_with_interrupt(delay, cls, None, None, None) delay *= retry_backoff continue diag = await _diagnose_connectivity() if not diag["internet_accessible"]: raise LocalNetworkError( "Unable to connect to the network. Please check your internet connection and try again." ) from e raise ApiServerError("The remote service appears unreachable at this time.") from e finally: if stop_evt is not None: stop_evt.set() if monitor_task: monitor_task.cancel() with contextlib.suppress(Exception): await monitor_task if req_task and not req_task.done(): req_task.cancel() with contextlib.suppress(Exception): await req_task if session: with contextlib.suppress(Exception): await session.close() if fhandle: with contextlib.suppress(Exception): fhandle.flush() fhandle.close() async def download_url_to_image_tensor( url: str, *, timeout: float = None, cls: type[COMFY_IO.ComfyNode] = None, ) -> torch.Tensor: """Downloads an image from a URL and returns a [B, H, W, C] tensor.""" result = BytesIO() await download_url_to_bytesio(url, result, timeout=timeout, cls=cls) return bytesio_to_image_tensor(result) async def download_url_to_video_output( video_url: str, *, timeout: float = None, max_retries: int = 5, cls: type[COMFY_IO.ComfyNode] = None, ) -> InputImpl.VideoFromFile: """Downloads a video from a URL and returns a `VIDEO` output.""" result = BytesIO() await download_url_to_bytesio(video_url, result, timeout=timeout, max_retries=max_retries, cls=cls) return InputImpl.VideoFromFile(result) async def download_url_as_bytesio( url: str, *, timeout: float = None, cls: type[COMFY_IO.ComfyNode] = None, ) -> BytesIO: """Downloads content from a URL and returns a new BytesIO (rewound to 0).""" result = BytesIO() await download_url_to_bytesio(url, result, timeout=timeout, cls=cls) return result def _generate_operation_id(method: str, url: str, attempt: int) -> str: try: parsed = urlparse(url) slug = (parsed.path.rsplit("/", 1)[-1] or parsed.netloc or "download").strip("/").replace("/", "_") except Exception: slug = "download" return f"{method}_{slug}_try{attempt}_{uuid.uuid4().hex[:8]}" async def download_url_to_file_3d( url: str, file_format: str, *, task_id: str | None = None, timeout: float | None = None, max_retries: int = 5, cls: type[COMFY_IO.ComfyNode] = None, ) -> Types.File3D: """Downloads a 3D model file from a URL into memory as BytesIO. If task_id is provided, also writes the file to disk in the output directory for backward compatibility with the old save-to-disk behavior. """ file_format = file_format.lstrip(".").lower() data = BytesIO() await download_url_to_bytesio( url, data, timeout=timeout, max_retries=max_retries, cls=cls, ) if task_id is not None: # This is only for backward compatability with current behavior when every 3D node is output node # All new API nodes should not use "task_id" and instead users should use "SaveGLB" node to save results output_dir = Path(get_output_directory()) output_path = output_dir / f"{task_id}.{file_format}" output_path.write_bytes(data.getvalue()) data.seek(0) return Types.File3D(source=data, file_format=file_format)
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api_nodes/util/download_helpers.py", "license": "GNU General Public License v3.0", "lines": 254, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_api_nodes/util/upload_helpers.py
import asyncio import contextlib import logging import time import uuid from io import BytesIO from urllib.parse import urlparse import aiohttp import torch from pydantic import BaseModel, Field from comfy_api.latest import IO, Input, Types from . import request_logger from ._helpers import is_processing_interrupted, sleep_with_interrupt from .client import ( ApiEndpoint, _diagnose_connectivity, _display_time_progress, sync_op, ) from .common_exceptions import ApiServerError, LocalNetworkError, ProcessingInterrupted from .conversions import ( audio_ndarray_to_bytesio, audio_tensor_to_contiguous_ndarray, tensor_to_bytesio, ) class UploadRequest(BaseModel): file_name: str = Field(..., description="Filename to upload") content_type: str | None = Field( None, description="Mime type of the file. For example: image/png, image/jpeg, video/mp4, etc.", ) class UploadResponse(BaseModel): download_url: str = Field(..., description="URL to GET uploaded file") upload_url: str = Field(..., description="URL to PUT file to upload") async def upload_images_to_comfyapi( cls: type[IO.ComfyNode], image: torch.Tensor | list[torch.Tensor], *, max_images: int = 8, mime_type: str | None = None, wait_label: str | None = "Uploading", show_batch_index: bool = True, total_pixels: int | None = 2048 * 2048, ) -> list[str]: """ Uploads images to ComfyUI API and returns download URLs. To upload multiple images, stack them in the batch dimension first. """ tensors: list[torch.Tensor] = [] if isinstance(image, list): for img in image: is_batch = len(img.shape) > 3 if is_batch: tensors.extend(img[i] for i in range(img.shape[0])) else: tensors.append(img) else: is_batch = len(image.shape) > 3 if is_batch: tensors.extend(image[i] for i in range(image.shape[0])) else: tensors.append(image) # if batched, try to upload each file if max_images is greater than 0 download_urls: list[str] = [] num_to_upload = min(len(tensors), max_images) batch_start_ts = time.monotonic() for idx in range(num_to_upload): tensor = tensors[idx] img_io = tensor_to_bytesio(tensor, total_pixels=total_pixels, mime_type=mime_type) effective_label = wait_label if wait_label and show_batch_index and num_to_upload > 1: effective_label = f"{wait_label} ({idx + 1}/{num_to_upload})" url = await upload_file_to_comfyapi(cls, img_io, img_io.name, mime_type, effective_label, batch_start_ts) download_urls.append(url) return download_urls async def upload_image_to_comfyapi( cls: type[IO.ComfyNode], image: torch.Tensor, *, mime_type: str | None = None, wait_label: str | None = "Uploading", total_pixels: int | None = 2048 * 2048, ) -> str: """Uploads a single image to ComfyUI API and returns its download URL.""" return ( await upload_images_to_comfyapi( cls, image, max_images=1, mime_type=mime_type, wait_label=wait_label, show_batch_index=False, total_pixels=total_pixels, ) )[0] async def upload_audio_to_comfyapi( cls: type[IO.ComfyNode], audio: Input.Audio, *, container_format: str = "mp4", codec_name: str = "aac", mime_type: str = "audio/mp4", ) -> str: """ Uploads a single audio input to ComfyUI API and returns its download URL. Encodes the raw waveform into the specified format before uploading. """ sample_rate: int = audio["sample_rate"] waveform: torch.Tensor = audio["waveform"] audio_data_np = audio_tensor_to_contiguous_ndarray(waveform) audio_bytes_io = audio_ndarray_to_bytesio(audio_data_np, sample_rate, container_format, codec_name) return await upload_file_to_comfyapi(cls, audio_bytes_io, f"{uuid.uuid4()}.{container_format}", mime_type) async def upload_video_to_comfyapi( cls: type[IO.ComfyNode], video: Input.Video, *, container: Types.VideoContainer = Types.VideoContainer.MP4, codec: Types.VideoCodec = Types.VideoCodec.H264, max_duration: int | None = None, wait_label: str | None = "Uploading", ) -> str: """ Uploads a single video to ComfyUI API and returns its download URL. Uses the specified container and codec for saving the video before upload. """ if max_duration is not None: try: actual_duration = video.get_duration() if actual_duration > max_duration: raise ValueError( f"Video duration ({actual_duration:.2f}s) exceeds the maximum allowed ({max_duration}s)." ) except Exception as e: logging.error("Error getting video duration: %s", str(e)) raise ValueError(f"Could not verify video duration from source: {e}") from e upload_mime_type = f"video/{container.value.lower()}" filename = f"{uuid.uuid4()}.{container.value.lower()}" # Convert VideoInput to BytesIO using specified container/codec video_bytes_io = BytesIO() video.save_to(video_bytes_io, format=container, codec=codec) video_bytes_io.seek(0) return await upload_file_to_comfyapi(cls, video_bytes_io, filename, upload_mime_type, wait_label) _3D_MIME_TYPES = { "glb": "model/gltf-binary", "obj": "model/obj", "fbx": "application/octet-stream", } async def upload_3d_model_to_comfyapi( cls: type[IO.ComfyNode], model_3d: Types.File3D, file_format: str, ) -> str: """Uploads a 3D model file to ComfyUI API and returns its download URL.""" return await upload_file_to_comfyapi( cls, model_3d.get_data(), f"{uuid.uuid4()}.{file_format}", _3D_MIME_TYPES.get(file_format, "application/octet-stream"), ) async def upload_file_to_comfyapi( cls: type[IO.ComfyNode], file_bytes_io: BytesIO, filename: str, upload_mime_type: str | None, wait_label: str | None = "Uploading", progress_origin_ts: float | None = None, ) -> str: """Uploads a single file to ComfyUI API and returns its download URL.""" if upload_mime_type is None: request_object = UploadRequest(file_name=filename) else: request_object = UploadRequest(file_name=filename, content_type=upload_mime_type) create_resp = await sync_op( cls, endpoint=ApiEndpoint(path="/customers/storage", method="POST"), data=request_object, response_model=UploadResponse, final_label_on_success=None, monitor_progress=False, ) await upload_file( cls, create_resp.upload_url, file_bytes_io, content_type=upload_mime_type, wait_label=wait_label, progress_origin_ts=progress_origin_ts, ) return create_resp.download_url async def upload_file( cls: type[IO.ComfyNode], upload_url: str, file: BytesIO | str, *, content_type: str | None = None, max_retries: int = 3, retry_delay: float = 1.0, retry_backoff: float = 2.0, wait_label: str | None = None, progress_origin_ts: float | None = None, ) -> None: """ Upload a file to a signed URL (e.g., S3 pre-signed PUT) with retries, Comfy progress display, and interruption. Raises: ProcessingInterrupted, LocalNetworkError, ApiServerError, Exception """ if isinstance(file, BytesIO): with contextlib.suppress(Exception): file.seek(0) data = file.read() elif isinstance(file, str): with open(file, "rb") as f: data = f.read() else: raise ValueError("file must be a BytesIO or a filesystem path string") headers: dict[str, str] = {} skip_auto_headers: set[str] = set() if content_type: headers["Content-Type"] = content_type else: skip_auto_headers.add("Content-Type") # Don't let aiohttp add Content-Type, it can break the signed request attempt = 0 delay = retry_delay start_ts = progress_origin_ts if progress_origin_ts is not None else time.monotonic() op_uuid = uuid.uuid4().hex[:8] while True: attempt += 1 operation_id = _generate_operation_id("PUT", upload_url, attempt, op_uuid) timeout = aiohttp.ClientTimeout(total=None) stop_evt = asyncio.Event() async def _monitor(): try: while not stop_evt.is_set(): if is_processing_interrupted(): return if wait_label: _display_time_progress(cls, wait_label, int(time.monotonic() - start_ts), None) await asyncio.sleep(1.0) except asyncio.CancelledError: return monitor_task = asyncio.create_task(_monitor()) sess: aiohttp.ClientSession | None = None try: request_logger.log_request_response( operation_id=operation_id, request_method="PUT", request_url=upload_url, request_headers=headers or None, request_params=None, request_data=f"[File data {len(data)} bytes]", ) sess = aiohttp.ClientSession(timeout=timeout) req = sess.put(upload_url, data=data, headers=headers, skip_auto_headers=skip_auto_headers) req_task = asyncio.create_task(req) done, pending = await asyncio.wait({req_task, monitor_task}, return_when=asyncio.FIRST_COMPLETED) if monitor_task in done and req_task in pending: req_task.cancel() raise ProcessingInterrupted("Upload cancelled") try: resp = await req_task except asyncio.CancelledError: raise ProcessingInterrupted("Upload cancelled") from None async with resp: if resp.status >= 400: with contextlib.suppress(Exception): try: body = await resp.json() except Exception: body = await resp.text() msg = f"Upload failed with status {resp.status}" request_logger.log_request_response( operation_id=operation_id, request_method="PUT", request_url=upload_url, response_status_code=resp.status, response_headers=dict(resp.headers), response_content=body, error_message=msg, ) if resp.status in {408, 429, 500, 502, 503, 504} and attempt <= max_retries: await sleep_with_interrupt( delay, cls, wait_label, start_ts, None, display_callback=_display_time_progress if wait_label else None, ) delay *= retry_backoff continue raise Exception(f"Failed to upload (HTTP {resp.status}).") request_logger.log_request_response( operation_id=operation_id, request_method="PUT", request_url=upload_url, response_status_code=resp.status, response_headers=dict(resp.headers), response_content="File uploaded successfully.", ) return except asyncio.CancelledError: raise ProcessingInterrupted("Task cancelled") from None except (aiohttp.ClientError, OSError) as e: if attempt <= max_retries: request_logger.log_request_response( operation_id=operation_id, request_method="PUT", request_url=upload_url, request_headers=headers or None, request_data=f"[File data {len(data)} bytes]", error_message=f"{type(e).__name__}: {str(e)} (will retry)", ) await sleep_with_interrupt( delay, cls, wait_label, start_ts, None, display_callback=_display_time_progress if wait_label else None, ) delay *= retry_backoff continue diag = await _diagnose_connectivity() if not diag["internet_accessible"]: raise LocalNetworkError( "Unable to connect to the network. Please check your internet connection and try again." ) from e raise ApiServerError("The API service appears unreachable at this time.") from e finally: stop_evt.set() if monitor_task: monitor_task.cancel() with contextlib.suppress(Exception): await monitor_task if sess: with contextlib.suppress(Exception): await sess.close() def _generate_operation_id(method: str, url: str, attempt: int, op_uuid: str) -> str: try: parsed = urlparse(url) slug = (parsed.path.rsplit("/", 1)[-1] or parsed.netloc or "upload").strip("/").replace("/", "_") except Exception: slug = "upload" return f"{method}_{slug}_{op_uuid}_try{attempt}"
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api_nodes/util/upload_helpers.py", "license": "GNU General Public License v3.0", "lines": 344, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy/nested_tensor.py
import torch class NestedTensor: def __init__(self, tensors): self.tensors = list(tensors) self.is_nested = True def _copy(self): return NestedTensor(self.tensors) def apply_operation(self, other, operation): o = self._copy() if isinstance(other, NestedTensor): for i, t in enumerate(o.tensors): o.tensors[i] = operation(t, other.tensors[i]) else: for i, t in enumerate(o.tensors): o.tensors[i] = operation(t, other) return o def __add__(self, b): return self.apply_operation(b, lambda x, y: x + y) def __sub__(self, b): return self.apply_operation(b, lambda x, y: x - y) def __mul__(self, b): return self.apply_operation(b, lambda x, y: x * y) # def __itruediv__(self, b): # return self.apply_operation(b, lambda x, y: x / y) def __truediv__(self, b): return self.apply_operation(b, lambda x, y: x / y) def __getitem__(self, *args, **kwargs): return self.apply_operation(None, lambda x, y: x.__getitem__(*args, **kwargs)) def unbind(self): return self.tensors def to(self, *args, **kwargs): o = self._copy() for i, t in enumerate(o.tensors): o.tensors[i] = t.to(*args, **kwargs) return o def new_ones(self, *args, **kwargs): return self.tensors[0].new_ones(*args, **kwargs) def float(self): return self.to(dtype=torch.float) def chunk(self, *args, **kwargs): return self.apply_operation(None, lambda x, y: x.chunk(*args, **kwargs)) def size(self): return self.tensors[0].size() @property def shape(self): return self.tensors[0].shape @property def ndim(self): dims = 0 for t in self.tensors: dims = max(t.ndim, dims) return dims @property def device(self): return self.tensors[0].device @property def dtype(self): return self.tensors[0].dtype @property def layout(self): return self.tensors[0].layout def cat_nested(tensors, *args, **kwargs): cated_tensors = [] for i in range(len(tensors[0].tensors)): tens = [] for j in range(len(tensors)): tens.append(tensors[j].tensors[i]) cated_tensors.append(torch.cat(tens, *args, **kwargs)) return NestedTensor(cated_tensors)
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/nested_tensor.py", "license": "GNU General Public License v3.0", "lines": 69, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:app/subgraph_manager.py
from __future__ import annotations from typing import TypedDict import os import folder_paths import glob from aiohttp import web import hashlib class Source: custom_node = "custom_node" templates = "templates" class SubgraphEntry(TypedDict): source: str """ Source of subgraph - custom_nodes vs templates. """ path: str """ Relative path of the subgraph file. For custom nodes, will be the relative directory like <custom_node_dir>/subgraphs/<name>.json """ name: str """ Name of subgraph file. """ info: CustomNodeSubgraphEntryInfo """ Additional info about subgraph; in the case of custom_nodes, will contain nodepack name """ data: str class CustomNodeSubgraphEntryInfo(TypedDict): node_pack: str """Node pack name.""" class SubgraphManager: def __init__(self): self.cached_custom_node_subgraphs: dict[SubgraphEntry] | None = None self.cached_blueprint_subgraphs: dict[SubgraphEntry] | None = None def _create_entry(self, file: str, source: str, node_pack: str) -> tuple[str, SubgraphEntry]: """Create a subgraph entry from a file path. Expects normalized path (forward slashes).""" entry_id = hashlib.sha256(f"{source}{file}".encode()).hexdigest() entry: SubgraphEntry = { "source": source, "name": os.path.splitext(os.path.basename(file))[0], "path": file, "info": {"node_pack": node_pack}, } return entry_id, entry async def load_entry_data(self, entry: SubgraphEntry): with open(entry['path'], 'r', encoding='utf-8') as f: entry['data'] = f.read() return entry async def sanitize_entry(self, entry: SubgraphEntry | None, remove_data=False) -> SubgraphEntry | None: if entry is None: return None entry = entry.copy() entry.pop('path', None) if remove_data: entry.pop('data', None) return entry async def sanitize_entries(self, entries: dict[str, SubgraphEntry], remove_data=False) -> dict[str, SubgraphEntry]: entries = entries.copy() for key in list(entries.keys()): entries[key] = await self.sanitize_entry(entries[key], remove_data) return entries async def get_custom_node_subgraphs(self, loadedModules, force_reload=False): """Load subgraphs from custom nodes.""" if not force_reload and self.cached_custom_node_subgraphs is not None: return self.cached_custom_node_subgraphs subgraphs_dict: dict[SubgraphEntry] = {} for folder in folder_paths.get_folder_paths("custom_nodes"): pattern = os.path.join(folder, "*/subgraphs/*.json") for file in glob.glob(pattern): file = file.replace('\\', '/') node_pack = "custom_nodes." + file.split('/')[-3] entry_id, entry = self._create_entry(file, Source.custom_node, node_pack) subgraphs_dict[entry_id] = entry self.cached_custom_node_subgraphs = subgraphs_dict return subgraphs_dict async def get_blueprint_subgraphs(self, force_reload=False): """Load subgraphs from the blueprints directory.""" if not force_reload and self.cached_blueprint_subgraphs is not None: return self.cached_blueprint_subgraphs subgraphs_dict: dict[SubgraphEntry] = {} blueprints_dir = os.path.join(os.path.dirname(os.path.dirname(__file__)), 'blueprints') if os.path.exists(blueprints_dir): for file in glob.glob(os.path.join(blueprints_dir, "*.json")): file = file.replace('\\', '/') entry_id, entry = self._create_entry(file, Source.templates, "comfyui") subgraphs_dict[entry_id] = entry self.cached_blueprint_subgraphs = subgraphs_dict return subgraphs_dict async def get_all_subgraphs(self, loadedModules, force_reload=False): """Get all subgraphs from all sources (custom nodes and blueprints).""" custom_node_subgraphs = await self.get_custom_node_subgraphs(loadedModules, force_reload) blueprint_subgraphs = await self.get_blueprint_subgraphs(force_reload) return {**custom_node_subgraphs, **blueprint_subgraphs} async def get_subgraph(self, id: str, loadedModules): """Get a specific subgraph by ID from any source.""" entry = (await self.get_all_subgraphs(loadedModules)).get(id) if entry is not None and entry.get('data') is None: await self.load_entry_data(entry) return entry def add_routes(self, routes, loadedModules): @routes.get("/global_subgraphs") async def get_global_subgraphs(request): subgraphs_dict = await self.get_all_subgraphs(loadedModules) return web.json_response(await self.sanitize_entries(subgraphs_dict, remove_data=True)) @routes.get("/global_subgraphs/{id}") async def get_global_subgraph(request): id = request.match_info.get("id", None) subgraph = await self.get_subgraph(id, loadedModules) return web.json_response(await self.sanitize_entry(subgraph))
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "app/subgraph_manager.py", "license": "GNU General Public License v3.0", "lines": 111, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy/ldm/mmaudio/vae/activations.py
# Implementation adapted from https://github.com/EdwardDixon/snake under the MIT license. # LICENSE is in incl_licenses directory. import torch from torch import nn, sin, pow from torch.nn import Parameter import comfy.model_management class Snake(nn.Module): ''' Implementation of a sine-based periodic activation function Shape: - Input: (B, C, T) - Output: (B, C, T), same shape as the input Parameters: - alpha - trainable parameter References: - This activation function is from this paper by Liu Ziyin, Tilman Hartwig, Masahito Ueda: https://arxiv.org/abs/2006.08195 Examples: >>> a1 = snake(256) >>> x = torch.randn(256) >>> x = a1(x) ''' def __init__(self, in_features, alpha=1.0, alpha_trainable=True, alpha_logscale=False): ''' Initialization. INPUT: - in_features: shape of the input - alpha: trainable parameter alpha is initialized to 1 by default, higher values = higher-frequency. alpha will be trained along with the rest of your model. ''' super(Snake, self).__init__() self.in_features = in_features # initialize alpha self.alpha_logscale = alpha_logscale if self.alpha_logscale: self.alpha = Parameter(torch.empty(in_features)) else: self.alpha = Parameter(torch.empty(in_features)) self.alpha.requires_grad = alpha_trainable self.no_div_by_zero = 0.000000001 def forward(self, x): ''' Forward pass of the function. Applies the function to the input elementwise. Snake ∶= x + 1/a * sin^2 (xa) ''' alpha = comfy.model_management.cast_to(self.alpha, dtype=x.dtype, device=x.device).unsqueeze(0).unsqueeze(-1) # line up with x to [B, C, T] if self.alpha_logscale: alpha = torch.exp(alpha) x = x + (1.0 / (alpha + self.no_div_by_zero)) * pow(sin(x * alpha), 2) return x class SnakeBeta(nn.Module): ''' A modified Snake function which uses separate parameters for the magnitude of the periodic components Shape: - Input: (B, C, T) - Output: (B, C, T), same shape as the input Parameters: - alpha - trainable parameter that controls frequency - beta - trainable parameter that controls magnitude References: - This activation function is a modified version based on this paper by Liu Ziyin, Tilman Hartwig, Masahito Ueda: https://arxiv.org/abs/2006.08195 Examples: >>> a1 = snakebeta(256) >>> x = torch.randn(256) >>> x = a1(x) ''' def __init__(self, in_features, alpha=1.0, alpha_trainable=True, alpha_logscale=False): ''' Initialization. INPUT: - in_features: shape of the input - alpha - trainable parameter that controls frequency - beta - trainable parameter that controls magnitude alpha is initialized to 1 by default, higher values = higher-frequency. beta is initialized to 1 by default, higher values = higher-magnitude. alpha will be trained along with the rest of your model. ''' super(SnakeBeta, self).__init__() self.in_features = in_features # initialize alpha self.alpha_logscale = alpha_logscale if self.alpha_logscale: self.alpha = Parameter(torch.empty(in_features)) self.beta = Parameter(torch.empty(in_features)) else: self.alpha = Parameter(torch.empty(in_features)) self.beta = Parameter(torch.empty(in_features)) self.alpha.requires_grad = alpha_trainable self.beta.requires_grad = alpha_trainable self.no_div_by_zero = 0.000000001 def forward(self, x): ''' Forward pass of the function. Applies the function to the input elementwise. SnakeBeta ∶= x + 1/b * sin^2 (xa) ''' alpha = comfy.model_management.cast_to(self.alpha, dtype=x.dtype, device=x.device).unsqueeze(0).unsqueeze(-1) # line up with x to [B, C, T] beta = comfy.model_management.cast_to(self.beta, dtype=x.dtype, device=x.device).unsqueeze(0).unsqueeze(-1) if self.alpha_logscale: alpha = torch.exp(alpha) beta = torch.exp(beta) x = x + (1.0 / (beta + self.no_div_by_zero)) * pow(sin(x * alpha), 2) return x
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/ldm/mmaudio/vae/activations.py", "license": "GNU General Public License v3.0", "lines": 106, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
documentation
Comfy-Org/ComfyUI:comfy/ldm/mmaudio/vae/alias_free_torch.py
import torch import torch.nn as nn import torch.nn.functional as F import math import comfy.model_management if 'sinc' in dir(torch): sinc = torch.sinc else: # This code is adopted from adefossez's julius.core.sinc under the MIT License # https://adefossez.github.io/julius/julius/core.html # LICENSE is in incl_licenses directory. def sinc(x: torch.Tensor): """ Implementation of sinc, i.e. sin(pi * x) / (pi * x) __Warning__: Different to julius.sinc, the input is multiplied by `pi`! """ return torch.where(x == 0, torch.tensor(1., device=x.device, dtype=x.dtype), torch.sin(math.pi * x) / math.pi / x) # This code is adopted from adefossez's julius.lowpass.LowPassFilters under the MIT License # https://adefossez.github.io/julius/julius/lowpass.html # LICENSE is in incl_licenses directory. def kaiser_sinc_filter1d(cutoff, half_width, kernel_size): # return filter [1,1,kernel_size] even = (kernel_size % 2 == 0) half_size = kernel_size // 2 #For kaiser window delta_f = 4 * half_width A = 2.285 * (half_size - 1) * math.pi * delta_f + 7.95 if A > 50.: beta = 0.1102 * (A - 8.7) elif A >= 21.: beta = 0.5842 * (A - 21)**0.4 + 0.07886 * (A - 21.) else: beta = 0. window = torch.kaiser_window(kernel_size, beta=beta, periodic=False) # ratio = 0.5/cutoff -> 2 * cutoff = 1 / ratio if even: time = (torch.arange(-half_size, half_size) + 0.5) else: time = torch.arange(kernel_size) - half_size if cutoff == 0: filter_ = torch.zeros_like(time) else: filter_ = 2 * cutoff * window * sinc(2 * cutoff * time) # Normalize filter to have sum = 1, otherwise we will have a small leakage # of the constant component in the input signal. filter_ /= filter_.sum() filter = filter_.view(1, 1, kernel_size) return filter class LowPassFilter1d(nn.Module): def __init__(self, cutoff=0.5, half_width=0.6, stride: int = 1, padding: bool = True, padding_mode: str = 'replicate', kernel_size: int = 12): # kernel_size should be even number for stylegan3 setup, # in this implementation, odd number is also possible. super().__init__() if cutoff < -0.: raise ValueError("Minimum cutoff must be larger than zero.") if cutoff > 0.5: raise ValueError("A cutoff above 0.5 does not make sense.") self.kernel_size = kernel_size self.even = (kernel_size % 2 == 0) self.pad_left = kernel_size // 2 - int(self.even) self.pad_right = kernel_size // 2 self.stride = stride self.padding = padding self.padding_mode = padding_mode filter = kaiser_sinc_filter1d(cutoff, half_width, kernel_size) self.register_buffer("filter", filter) #input [B, C, T] def forward(self, x): _, C, _ = x.shape if self.padding: x = F.pad(x, (self.pad_left, self.pad_right), mode=self.padding_mode) out = F.conv1d(x, comfy.model_management.cast_to(self.filter.expand(C, -1, -1), dtype=x.dtype, device=x.device), stride=self.stride, groups=C) return out class UpSample1d(nn.Module): def __init__(self, ratio=2, kernel_size=None): super().__init__() self.ratio = ratio self.kernel_size = int(6 * ratio // 2) * 2 if kernel_size is None else kernel_size self.stride = ratio self.pad = self.kernel_size // ratio - 1 self.pad_left = self.pad * self.stride + (self.kernel_size - self.stride) // 2 self.pad_right = self.pad * self.stride + (self.kernel_size - self.stride + 1) // 2 filter = kaiser_sinc_filter1d(cutoff=0.5 / ratio, half_width=0.6 / ratio, kernel_size=self.kernel_size) self.register_buffer("filter", filter) # x: [B, C, T] def forward(self, x): _, C, _ = x.shape x = F.pad(x, (self.pad, self.pad), mode='replicate') x = self.ratio * F.conv_transpose1d( x, comfy.model_management.cast_to(self.filter.expand(C, -1, -1), dtype=x.dtype, device=x.device), stride=self.stride, groups=C) x = x[..., self.pad_left:-self.pad_right] return x class DownSample1d(nn.Module): def __init__(self, ratio=2, kernel_size=None): super().__init__() self.ratio = ratio self.kernel_size = int(6 * ratio // 2) * 2 if kernel_size is None else kernel_size self.lowpass = LowPassFilter1d(cutoff=0.5 / ratio, half_width=0.6 / ratio, stride=ratio, kernel_size=self.kernel_size) def forward(self, x): xx = self.lowpass(x) return xx class Activation1d(nn.Module): def __init__(self, activation, up_ratio: int = 2, down_ratio: int = 2, up_kernel_size: int = 12, down_kernel_size: int = 12): super().__init__() self.up_ratio = up_ratio self.down_ratio = down_ratio self.act = activation self.upsample = UpSample1d(up_ratio, up_kernel_size) self.downsample = DownSample1d(down_ratio, down_kernel_size) # x: [B,C,T] def forward(self, x): x = self.upsample(x) x = self.act(x) x = self.downsample(x) return x
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/ldm/mmaudio/vae/alias_free_torch.py", "license": "GNU General Public License v3.0", "lines": 134, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy/ldm/mmaudio/vae/autoencoder.py
from typing import Literal import torch import torch.nn as nn from .distributions import DiagonalGaussianDistribution from .vae import VAE_16k from .bigvgan import BigVGANVocoder import logging try: import torchaudio except: logging.warning("torchaudio missing, MMAudio VAE model will be broken") def dynamic_range_compression_torch(x, C=1, clip_val=1e-5, *, norm_fn): return norm_fn(torch.clamp(x, min=clip_val) * C) def spectral_normalize_torch(magnitudes, norm_fn): output = dynamic_range_compression_torch(magnitudes, norm_fn=norm_fn) return output class MelConverter(nn.Module): def __init__( self, *, sampling_rate: float, n_fft: int, num_mels: int, hop_size: int, win_size: int, fmin: float, fmax: float, norm_fn, ): super().__init__() self.sampling_rate = sampling_rate self.n_fft = n_fft self.num_mels = num_mels self.hop_size = hop_size self.win_size = win_size self.fmin = fmin self.fmax = fmax self.norm_fn = norm_fn # mel = librosa_mel_fn(sr=self.sampling_rate, # n_fft=self.n_fft, # n_mels=self.num_mels, # fmin=self.fmin, # fmax=self.fmax) # mel_basis = torch.from_numpy(mel).float() mel_basis = torch.empty((num_mels, 1 + n_fft // 2)) hann_window = torch.hann_window(self.win_size) self.register_buffer('mel_basis', mel_basis) self.register_buffer('hann_window', hann_window) @property def device(self): return self.mel_basis.device def forward(self, waveform: torch.Tensor, center: bool = False) -> torch.Tensor: waveform = waveform.clamp(min=-1., max=1.).to(self.device) waveform = torch.nn.functional.pad( waveform.unsqueeze(1), [int((self.n_fft - self.hop_size) / 2), int((self.n_fft - self.hop_size) / 2)], mode='reflect') waveform = waveform.squeeze(1) spec = torch.stft(waveform, self.n_fft, hop_length=self.hop_size, win_length=self.win_size, window=self.hann_window, center=center, pad_mode='reflect', normalized=False, onesided=True, return_complex=True) spec = torch.view_as_real(spec) spec = torch.sqrt(spec.pow(2).sum(-1) + (1e-9)) spec = torch.matmul(self.mel_basis, spec) spec = spectral_normalize_torch(spec, self.norm_fn) return spec class AudioAutoencoder(nn.Module): def __init__( self, *, # ckpt_path: str, mode=Literal['16k', '44k'], need_vae_encoder: bool = True, ): super().__init__() assert mode == "16k", "Only 16k mode is supported currently." self.mel_converter = MelConverter(sampling_rate=16_000, n_fft=1024, num_mels=80, hop_size=256, win_size=1024, fmin=0, fmax=8_000, norm_fn=torch.log10) self.vae = VAE_16k().eval() bigvgan_config = { "resblock": "1", "num_mels": 80, "upsample_rates": [4, 4, 2, 2, 2, 2], "upsample_kernel_sizes": [8, 8, 4, 4, 4, 4], "upsample_initial_channel": 1536, "resblock_kernel_sizes": [3, 7, 11], "resblock_dilation_sizes": [ [1, 3, 5], [1, 3, 5], [1, 3, 5], ], "activation": "snakebeta", "snake_logscale": True, } self.vocoder = BigVGANVocoder( bigvgan_config ).eval() @torch.inference_mode() def encode_audio(self, x) -> DiagonalGaussianDistribution: # x: (B * L) mel = self.mel_converter(x) dist = self.vae.encode(mel) return dist @torch.no_grad() def decode(self, z): mel_decoded = self.vae.decode(z) audio = self.vocoder(mel_decoded) audio = torchaudio.functional.resample(audio, 16000, 44100) return audio @torch.no_grad() def encode(self, audio): audio = audio.mean(dim=1) audio = torchaudio.functional.resample(audio, 44100, 16000) dist = self.encode_audio(audio) return dist.mean
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/ldm/mmaudio/vae/autoencoder.py", "license": "GNU General Public License v3.0", "lines": 129, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy/ldm/mmaudio/vae/bigvgan.py
# Copyright (c) 2022 NVIDIA CORPORATION. # Licensed under the MIT license. # Adapted from https://github.com/jik876/hifi-gan under the MIT license. # LICENSE is in incl_licenses directory. import torch import torch.nn as nn from types import SimpleNamespace from . import activations from .alias_free_torch import Activation1d import comfy.ops ops = comfy.ops.disable_weight_init def get_padding(kernel_size, dilation=1): return int((kernel_size * dilation - dilation) / 2) class AMPBlock1(torch.nn.Module): def __init__(self, h, channels, kernel_size=3, dilation=(1, 3, 5), activation=None): super(AMPBlock1, self).__init__() self.h = h self.convs1 = nn.ModuleList([ ops.Conv1d(channels, channels, kernel_size, 1, dilation=dilation[0], padding=get_padding(kernel_size, dilation[0])), ops.Conv1d(channels, channels, kernel_size, 1, dilation=dilation[1], padding=get_padding(kernel_size, dilation[1])), ops.Conv1d(channels, channels, kernel_size, 1, dilation=dilation[2], padding=get_padding(kernel_size, dilation[2])) ]) self.convs2 = nn.ModuleList([ ops.Conv1d(channels, channels, kernel_size, 1, dilation=1, padding=get_padding(kernel_size, 1)), ops.Conv1d(channels, channels, kernel_size, 1, dilation=1, padding=get_padding(kernel_size, 1)), ops.Conv1d(channels, channels, kernel_size, 1, dilation=1, padding=get_padding(kernel_size, 1)) ]) self.num_layers = len(self.convs1) + len(self.convs2) # total number of conv layers if activation == 'snake': # periodic nonlinearity with snake function and anti-aliasing self.activations = nn.ModuleList([ Activation1d( activation=activations.Snake(channels, alpha_logscale=h.snake_logscale)) for _ in range(self.num_layers) ]) elif activation == 'snakebeta': # periodic nonlinearity with snakebeta function and anti-aliasing self.activations = nn.ModuleList([ Activation1d( activation=activations.SnakeBeta(channels, alpha_logscale=h.snake_logscale)) for _ in range(self.num_layers) ]) else: raise NotImplementedError( "activation incorrectly specified. check the config file and look for 'activation'." ) def forward(self, x): acts1, acts2 = self.activations[::2], self.activations[1::2] for c1, c2, a1, a2 in zip(self.convs1, self.convs2, acts1, acts2): xt = a1(x) xt = c1(xt) xt = a2(xt) xt = c2(xt) x = xt + x return x class AMPBlock2(torch.nn.Module): def __init__(self, h, channels, kernel_size=3, dilation=(1, 3), activation=None): super(AMPBlock2, self).__init__() self.h = h self.convs = nn.ModuleList([ ops.Conv1d(channels, channels, kernel_size, 1, dilation=dilation[0], padding=get_padding(kernel_size, dilation[0])), ops.Conv1d(channels, channels, kernel_size, 1, dilation=dilation[1], padding=get_padding(kernel_size, dilation[1])) ]) self.num_layers = len(self.convs) # total number of conv layers if activation == 'snake': # periodic nonlinearity with snake function and anti-aliasing self.activations = nn.ModuleList([ Activation1d( activation=activations.Snake(channels, alpha_logscale=h.snake_logscale)) for _ in range(self.num_layers) ]) elif activation == 'snakebeta': # periodic nonlinearity with snakebeta function and anti-aliasing self.activations = nn.ModuleList([ Activation1d( activation=activations.SnakeBeta(channels, alpha_logscale=h.snake_logscale)) for _ in range(self.num_layers) ]) else: raise NotImplementedError( "activation incorrectly specified. check the config file and look for 'activation'." ) def forward(self, x): for c, a in zip(self.convs, self.activations): xt = a(x) xt = c(xt) x = xt + x return x class BigVGANVocoder(torch.nn.Module): # this is our main BigVGAN model. Applies anti-aliased periodic activation for resblocks. def __init__(self, h): super().__init__() if isinstance(h, dict): h = SimpleNamespace(**h) self.h = h self.num_kernels = len(h.resblock_kernel_sizes) self.num_upsamples = len(h.upsample_rates) # pre conv self.conv_pre = ops.Conv1d(h.num_mels, h.upsample_initial_channel, 7, 1, padding=3) # define which AMPBlock to use. BigVGAN uses AMPBlock1 as default resblock = AMPBlock1 if h.resblock == '1' else AMPBlock2 # transposed conv-based upsamplers. does not apply anti-aliasing self.ups = nn.ModuleList() for i, (u, k) in enumerate(zip(h.upsample_rates, h.upsample_kernel_sizes)): self.ups.append( nn.ModuleList([ ops.ConvTranspose1d(h.upsample_initial_channel // (2**i), h.upsample_initial_channel // (2**(i + 1)), k, u, padding=(k - u) // 2) ])) # residual blocks using anti-aliased multi-periodicity composition modules (AMP) self.resblocks = nn.ModuleList() for i in range(len(self.ups)): ch = h.upsample_initial_channel // (2**(i + 1)) for j, (k, d) in enumerate(zip(h.resblock_kernel_sizes, h.resblock_dilation_sizes)): self.resblocks.append(resblock(h, ch, k, d, activation=h.activation)) # post conv if h.activation == "snake": # periodic nonlinearity with snake function and anti-aliasing activation_post = activations.Snake(ch, alpha_logscale=h.snake_logscale) self.activation_post = Activation1d(activation=activation_post) elif h.activation == "snakebeta": # periodic nonlinearity with snakebeta function and anti-aliasing activation_post = activations.SnakeBeta(ch, alpha_logscale=h.snake_logscale) self.activation_post = Activation1d(activation=activation_post) else: raise NotImplementedError( "activation incorrectly specified. check the config file and look for 'activation'." ) self.conv_post = ops.Conv1d(ch, 1, 7, 1, padding=3) def forward(self, x): # pre conv x = self.conv_pre(x) for i in range(self.num_upsamples): # upsampling for i_up in range(len(self.ups[i])): x = self.ups[i][i_up](x) # AMP blocks xs = None for j in range(self.num_kernels): if xs is None: xs = self.resblocks[i * self.num_kernels + j](x) else: xs += self.resblocks[i * self.num_kernels + j](x) x = xs / self.num_kernels # post conv x = self.activation_post(x) x = self.conv_post(x) x = torch.tanh(x) return x
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/ldm/mmaudio/vae/bigvgan.py", "license": "GNU General Public License v3.0", "lines": 186, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
Comfy-Org/ComfyUI:comfy/ldm/mmaudio/vae/distributions.py
import torch import numpy as np class AbstractDistribution: def sample(self): raise NotImplementedError() def mode(self): raise NotImplementedError() class DiracDistribution(AbstractDistribution): def __init__(self, value): self.value = value def sample(self): return self.value def mode(self): return self.value class DiagonalGaussianDistribution(object): def __init__(self, parameters, deterministic=False): self.parameters = parameters self.mean, self.logvar = torch.chunk(parameters, 2, dim=1) self.logvar = torch.clamp(self.logvar, -30.0, 20.0) self.deterministic = deterministic self.std = torch.exp(0.5 * self.logvar) self.var = torch.exp(self.logvar) if self.deterministic: self.var = self.std = torch.zeros_like(self.mean, device=self.parameters.device) def sample(self): x = self.mean + self.std * torch.randn(self.mean.shape, device=self.parameters.device) return x def kl(self, other=None): if self.deterministic: return torch.Tensor([0.]) else: if other is None: return 0.5 * torch.sum(torch.pow(self.mean, 2) + self.var - 1.0 - self.logvar, dim=[1, 2, 3]) else: return 0.5 * torch.sum( torch.pow(self.mean - other.mean, 2) / other.var + self.var / other.var - 1.0 - self.logvar + other.logvar, dim=[1, 2, 3]) def nll(self, sample, dims=[1,2,3]): if self.deterministic: return torch.Tensor([0.]) logtwopi = np.log(2.0 * np.pi) return 0.5 * torch.sum( logtwopi + self.logvar + torch.pow(sample - self.mean, 2) / self.var, dim=dims) def mode(self): return self.mean def normal_kl(mean1, logvar1, mean2, logvar2): """ source: https://github.com/openai/guided-diffusion/blob/27c20a8fab9cb472df5d6bdd6c8d11c8f430b924/guided_diffusion/losses.py#L12 Compute the KL divergence between two gaussians. Shapes are automatically broadcasted, so batches can be compared to scalars, among other use cases. """ tensor = None for obj in (mean1, logvar1, mean2, logvar2): if isinstance(obj, torch.Tensor): tensor = obj break assert tensor is not None, "at least one argument must be a Tensor" # Force variances to be Tensors. Broadcasting helps convert scalars to # Tensors, but it does not work for torch.exp(). logvar1, logvar2 = [ x if isinstance(x, torch.Tensor) else torch.tensor(x).to(tensor) for x in (logvar1, logvar2) ] return 0.5 * ( -1.0 + logvar2 - logvar1 + torch.exp(logvar1 - logvar2) + ((mean1 - mean2) ** 2) * torch.exp(-logvar2) )
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/ldm/mmaudio/vae/distributions.py", "license": "GNU General Public License v3.0", "lines": 75, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy/ldm/mmaudio/vae/vae.py
import logging from typing import Optional import torch import torch.nn as nn from .vae_modules import (AttnBlock1D, Downsample1D, ResnetBlock1D, Upsample1D, nonlinearity) from .distributions import DiagonalGaussianDistribution import comfy.ops ops = comfy.ops.disable_weight_init log = logging.getLogger() DATA_MEAN_80D = [ -1.6058, -1.3676, -1.2520, -1.2453, -1.2078, -1.2224, -1.2419, -1.2439, -1.2922, -1.2927, -1.3170, -1.3543, -1.3401, -1.3836, -1.3907, -1.3912, -1.4313, -1.4152, -1.4527, -1.4728, -1.4568, -1.5101, -1.5051, -1.5172, -1.5623, -1.5373, -1.5746, -1.5687, -1.6032, -1.6131, -1.6081, -1.6331, -1.6489, -1.6489, -1.6700, -1.6738, -1.6953, -1.6969, -1.7048, -1.7280, -1.7361, -1.7495, -1.7658, -1.7814, -1.7889, -1.8064, -1.8221, -1.8377, -1.8417, -1.8643, -1.8857, -1.8929, -1.9173, -1.9379, -1.9531, -1.9673, -1.9824, -2.0042, -2.0215, -2.0436, -2.0766, -2.1064, -2.1418, -2.1855, -2.2319, -2.2767, -2.3161, -2.3572, -2.3954, -2.4282, -2.4659, -2.5072, -2.5552, -2.6074, -2.6584, -2.7107, -2.7634, -2.8266, -2.8981, -2.9673 ] DATA_STD_80D = [ 1.0291, 1.0411, 1.0043, 0.9820, 0.9677, 0.9543, 0.9450, 0.9392, 0.9343, 0.9297, 0.9276, 0.9263, 0.9242, 0.9254, 0.9232, 0.9281, 0.9263, 0.9315, 0.9274, 0.9247, 0.9277, 0.9199, 0.9188, 0.9194, 0.9160, 0.9161, 0.9146, 0.9161, 0.9100, 0.9095, 0.9145, 0.9076, 0.9066, 0.9095, 0.9032, 0.9043, 0.9038, 0.9011, 0.9019, 0.9010, 0.8984, 0.8983, 0.8986, 0.8961, 0.8962, 0.8978, 0.8962, 0.8973, 0.8993, 0.8976, 0.8995, 0.9016, 0.8982, 0.8972, 0.8974, 0.8949, 0.8940, 0.8947, 0.8936, 0.8939, 0.8951, 0.8956, 0.9017, 0.9167, 0.9436, 0.9690, 1.0003, 1.0225, 1.0381, 1.0491, 1.0545, 1.0604, 1.0761, 1.0929, 1.1089, 1.1196, 1.1176, 1.1156, 1.1117, 1.1070 ] DATA_MEAN_128D = [ -3.3462, -2.6723, -2.4893, -2.3143, -2.2664, -2.3317, -2.1802, -2.4006, -2.2357, -2.4597, -2.3717, -2.4690, -2.5142, -2.4919, -2.6610, -2.5047, -2.7483, -2.5926, -2.7462, -2.7033, -2.7386, -2.8112, -2.7502, -2.9594, -2.7473, -3.0035, -2.8891, -2.9922, -2.9856, -3.0157, -3.1191, -2.9893, -3.1718, -3.0745, -3.1879, -3.2310, -3.1424, -3.2296, -3.2791, -3.2782, -3.2756, -3.3134, -3.3509, -3.3750, -3.3951, -3.3698, -3.4505, -3.4509, -3.5089, -3.4647, -3.5536, -3.5788, -3.5867, -3.6036, -3.6400, -3.6747, -3.7072, -3.7279, -3.7283, -3.7795, -3.8259, -3.8447, -3.8663, -3.9182, -3.9605, -3.9861, -4.0105, -4.0373, -4.0762, -4.1121, -4.1488, -4.1874, -4.2461, -4.3170, -4.3639, -4.4452, -4.5282, -4.6297, -4.7019, -4.7960, -4.8700, -4.9507, -5.0303, -5.0866, -5.1634, -5.2342, -5.3242, -5.4053, -5.4927, -5.5712, -5.6464, -5.7052, -5.7619, -5.8410, -5.9188, -6.0103, -6.0955, -6.1673, -6.2362, -6.3120, -6.3926, -6.4797, -6.5565, -6.6511, -6.8130, -6.9961, -7.1275, -7.2457, -7.3576, -7.4663, -7.6136, -7.7469, -7.8815, -8.0132, -8.1515, -8.3071, -8.4722, -8.7418, -9.3975, -9.6628, -9.7671, -9.8863, -9.9992, -10.0860, -10.1709, -10.5418, -11.2795, -11.3861 ] DATA_STD_128D = [ 2.3804, 2.4368, 2.3772, 2.3145, 2.2803, 2.2510, 2.2316, 2.2083, 2.1996, 2.1835, 2.1769, 2.1659, 2.1631, 2.1618, 2.1540, 2.1606, 2.1571, 2.1567, 2.1612, 2.1579, 2.1679, 2.1683, 2.1634, 2.1557, 2.1668, 2.1518, 2.1415, 2.1449, 2.1406, 2.1350, 2.1313, 2.1415, 2.1281, 2.1352, 2.1219, 2.1182, 2.1327, 2.1195, 2.1137, 2.1080, 2.1179, 2.1036, 2.1087, 2.1036, 2.1015, 2.1068, 2.0975, 2.0991, 2.0902, 2.1015, 2.0857, 2.0920, 2.0893, 2.0897, 2.0910, 2.0881, 2.0925, 2.0873, 2.0960, 2.0900, 2.0957, 2.0958, 2.0978, 2.0936, 2.0886, 2.0905, 2.0845, 2.0855, 2.0796, 2.0840, 2.0813, 2.0817, 2.0838, 2.0840, 2.0917, 2.1061, 2.1431, 2.1976, 2.2482, 2.3055, 2.3700, 2.4088, 2.4372, 2.4609, 2.4731, 2.4847, 2.5072, 2.5451, 2.5772, 2.6147, 2.6529, 2.6596, 2.6645, 2.6726, 2.6803, 2.6812, 2.6899, 2.6916, 2.6931, 2.6998, 2.7062, 2.7262, 2.7222, 2.7158, 2.7041, 2.7485, 2.7491, 2.7451, 2.7485, 2.7233, 2.7297, 2.7233, 2.7145, 2.6958, 2.6788, 2.6439, 2.6007, 2.4786, 2.2469, 2.1877, 2.1392, 2.0717, 2.0107, 1.9676, 1.9140, 1.7102, 0.9101, 0.7164 ] class VAE(nn.Module): def __init__( self, *, data_dim: int, embed_dim: int, hidden_dim: int, ): super().__init__() if data_dim == 80: self.data_mean = nn.Buffer(torch.tensor(DATA_MEAN_80D, dtype=torch.float32)) self.data_std = nn.Buffer(torch.tensor(DATA_STD_80D, dtype=torch.float32)) elif data_dim == 128: self.data_mean = nn.Buffer(torch.tensor(DATA_MEAN_128D, dtype=torch.float32)) self.data_std = nn.Buffer(torch.tensor(DATA_STD_128D, dtype=torch.float32)) self.data_mean = self.data_mean.view(1, -1, 1) self.data_std = self.data_std.view(1, -1, 1) self.encoder = Encoder1D( dim=hidden_dim, ch_mult=(1, 2, 4), num_res_blocks=2, attn_layers=[3], down_layers=[0], in_dim=data_dim, embed_dim=embed_dim, ) self.decoder = Decoder1D( dim=hidden_dim, ch_mult=(1, 2, 4), num_res_blocks=2, attn_layers=[3], down_layers=[0], in_dim=data_dim, out_dim=data_dim, embed_dim=embed_dim, ) self.embed_dim = embed_dim # self.quant_conv = nn.Conv1d(2 * embed_dim, 2 * embed_dim, 1) # self.post_quant_conv = nn.Conv1d(embed_dim, embed_dim, 1) self.initialize_weights() def initialize_weights(self): pass def encode(self, x: torch.Tensor, normalize: bool = True) -> DiagonalGaussianDistribution: if normalize: x = self.normalize(x) moments = self.encoder(x) posterior = DiagonalGaussianDistribution(moments) return posterior def decode(self, z: torch.Tensor, unnormalize: bool = True) -> torch.Tensor: dec = self.decoder(z) if unnormalize: dec = self.unnormalize(dec) return dec def normalize(self, x: torch.Tensor) -> torch.Tensor: return (x - comfy.model_management.cast_to(self.data_mean, dtype=x.dtype, device=x.device)) / comfy.model_management.cast_to(self.data_std, dtype=x.dtype, device=x.device) def unnormalize(self, x: torch.Tensor) -> torch.Tensor: return x * comfy.model_management.cast_to(self.data_std, dtype=x.dtype, device=x.device) + comfy.model_management.cast_to(self.data_mean, dtype=x.dtype, device=x.device) def forward( self, x: torch.Tensor, sample_posterior: bool = True, rng: Optional[torch.Generator] = None, normalize: bool = True, unnormalize: bool = True, ) -> tuple[torch.Tensor, DiagonalGaussianDistribution]: posterior = self.encode(x, normalize=normalize) if sample_posterior: z = posterior.sample(rng) else: z = posterior.mode() dec = self.decode(z, unnormalize=unnormalize) return dec, posterior def load_weights(self, src_dict) -> None: self.load_state_dict(src_dict, strict=True) @property def device(self) -> torch.device: return next(self.parameters()).device def get_last_layer(self): return self.decoder.conv_out.weight def remove_weight_norm(self): return self class Encoder1D(nn.Module): def __init__(self, *, dim: int, ch_mult: tuple[int] = (1, 2, 4, 8), num_res_blocks: int, attn_layers: list[int] = [], down_layers: list[int] = [], resamp_with_conv: bool = True, in_dim: int, embed_dim: int, double_z: bool = True, kernel_size: int = 3, clip_act: float = 256.0): super().__init__() self.dim = dim self.num_layers = len(ch_mult) self.num_res_blocks = num_res_blocks self.in_channels = in_dim self.clip_act = clip_act self.down_layers = down_layers self.attn_layers = attn_layers self.conv_in = ops.Conv1d(in_dim, self.dim, kernel_size=kernel_size, padding=kernel_size // 2, bias=False) in_ch_mult = (1, ) + tuple(ch_mult) self.in_ch_mult = in_ch_mult # downsampling self.down = nn.ModuleList() for i_level in range(self.num_layers): block = nn.ModuleList() attn = nn.ModuleList() block_in = dim * in_ch_mult[i_level] block_out = dim * ch_mult[i_level] for i_block in range(self.num_res_blocks): block.append( ResnetBlock1D(in_dim=block_in, out_dim=block_out, kernel_size=kernel_size, use_norm=True)) block_in = block_out if i_level in attn_layers: attn.append(AttnBlock1D(block_in)) down = nn.Module() down.block = block down.attn = attn if i_level in down_layers: down.downsample = Downsample1D(block_in, resamp_with_conv) self.down.append(down) # middle self.mid = nn.Module() self.mid.block_1 = ResnetBlock1D(in_dim=block_in, out_dim=block_in, kernel_size=kernel_size, use_norm=True) self.mid.attn_1 = AttnBlock1D(block_in) self.mid.block_2 = ResnetBlock1D(in_dim=block_in, out_dim=block_in, kernel_size=kernel_size, use_norm=True) # end self.conv_out = ops.Conv1d(block_in, 2 * embed_dim if double_z else embed_dim, kernel_size=kernel_size, padding=kernel_size // 2, bias=False) self.learnable_gain = nn.Parameter(torch.zeros([])) def forward(self, x): # downsampling h = self.conv_in(x) for i_level in range(self.num_layers): for i_block in range(self.num_res_blocks): h = self.down[i_level].block[i_block](h) if len(self.down[i_level].attn) > 0: h = self.down[i_level].attn[i_block](h) h = h.clamp(-self.clip_act, self.clip_act) if i_level in self.down_layers: h = self.down[i_level].downsample(h) # middle h = self.mid.block_1(h) h = self.mid.attn_1(h) h = self.mid.block_2(h) h = h.clamp(-self.clip_act, self.clip_act) # end h = nonlinearity(h) h = self.conv_out(h) * (self.learnable_gain + 1) return h class Decoder1D(nn.Module): def __init__(self, *, dim: int, out_dim: int, ch_mult: tuple[int] = (1, 2, 4, 8), num_res_blocks: int, attn_layers: list[int] = [], down_layers: list[int] = [], kernel_size: int = 3, resamp_with_conv: bool = True, in_dim: int, embed_dim: int, clip_act: float = 256.0): super().__init__() self.ch = dim self.num_layers = len(ch_mult) self.num_res_blocks = num_res_blocks self.in_channels = in_dim self.clip_act = clip_act self.down_layers = [i + 1 for i in down_layers] # each downlayer add one # compute in_ch_mult, block_in and curr_res at lowest res block_in = dim * ch_mult[self.num_layers - 1] # z to block_in self.conv_in = ops.Conv1d(embed_dim, block_in, kernel_size=kernel_size, padding=kernel_size // 2, bias=False) # middle self.mid = nn.Module() self.mid.block_1 = ResnetBlock1D(in_dim=block_in, out_dim=block_in, use_norm=True) self.mid.attn_1 = AttnBlock1D(block_in) self.mid.block_2 = ResnetBlock1D(in_dim=block_in, out_dim=block_in, use_norm=True) # upsampling self.up = nn.ModuleList() for i_level in reversed(range(self.num_layers)): block = nn.ModuleList() attn = nn.ModuleList() block_out = dim * ch_mult[i_level] for i_block in range(self.num_res_blocks + 1): block.append(ResnetBlock1D(in_dim=block_in, out_dim=block_out, use_norm=True)) block_in = block_out if i_level in attn_layers: attn.append(AttnBlock1D(block_in)) up = nn.Module() up.block = block up.attn = attn if i_level in self.down_layers: up.upsample = Upsample1D(block_in, resamp_with_conv) self.up.insert(0, up) # prepend to get consistent order # end self.conv_out = ops.Conv1d(block_in, out_dim, kernel_size=kernel_size, padding=kernel_size // 2, bias=False) self.learnable_gain = nn.Parameter(torch.zeros([])) def forward(self, z): # z to block_in h = self.conv_in(z) # middle h = self.mid.block_1(h) h = self.mid.attn_1(h) h = self.mid.block_2(h) h = h.clamp(-self.clip_act, self.clip_act) # upsampling for i_level in reversed(range(self.num_layers)): for i_block in range(self.num_res_blocks + 1): h = self.up[i_level].block[i_block](h) if len(self.up[i_level].attn) > 0: h = self.up[i_level].attn[i_block](h) h = h.clamp(-self.clip_act, self.clip_act) if i_level in self.down_layers: h = self.up[i_level].upsample(h) h = nonlinearity(h) h = self.conv_out(h) * (self.learnable_gain + 1) return h def VAE_16k(**kwargs) -> VAE: return VAE(data_dim=80, embed_dim=20, hidden_dim=384, **kwargs) def VAE_44k(**kwargs) -> VAE: return VAE(data_dim=128, embed_dim=40, hidden_dim=512, **kwargs) def get_my_vae(name: str, **kwargs) -> VAE: if name == '16k': return VAE_16k(**kwargs) if name == '44k': return VAE_44k(**kwargs) raise ValueError(f'Unknown model: {name}')
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/ldm/mmaudio/vae/vae.py", "license": "GNU General Public License v3.0", "lines": 301, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy/ldm/mmaudio/vae/vae_modules.py
import torch import torch.nn as nn import torch.nn.functional as F from comfy.ldm.modules.diffusionmodules.model import vae_attention import math import comfy.ops ops = comfy.ops.disable_weight_init def nonlinearity(x): # swish return torch.nn.functional.silu(x) / 0.596 def mp_sum(a, b, t=0.5): return a.lerp(b, t) / math.sqrt((1 - t)**2 + t**2) def normalize(x, dim=None, eps=1e-4): if dim is None: dim = list(range(1, x.ndim)) norm = torch.linalg.vector_norm(x, dim=dim, keepdim=True, dtype=torch.float32) norm = torch.add(eps, norm, alpha=math.sqrt(norm.numel() / x.numel())) return x / norm.to(x.dtype) class ResnetBlock1D(nn.Module): def __init__(self, *, in_dim, out_dim=None, conv_shortcut=False, kernel_size=3, use_norm=True): super().__init__() self.in_dim = in_dim out_dim = in_dim if out_dim is None else out_dim self.out_dim = out_dim self.use_conv_shortcut = conv_shortcut self.use_norm = use_norm self.conv1 = ops.Conv1d(in_dim, out_dim, kernel_size=kernel_size, padding=kernel_size // 2, bias=False) self.conv2 = ops.Conv1d(out_dim, out_dim, kernel_size=kernel_size, padding=kernel_size // 2, bias=False) if self.in_dim != self.out_dim: if self.use_conv_shortcut: self.conv_shortcut = ops.Conv1d(in_dim, out_dim, kernel_size=kernel_size, padding=kernel_size // 2, bias=False) else: self.nin_shortcut = ops.Conv1d(in_dim, out_dim, kernel_size=1, padding=0, bias=False) def forward(self, x: torch.Tensor) -> torch.Tensor: # pixel norm if self.use_norm: x = normalize(x, dim=1) h = x h = nonlinearity(h) h = self.conv1(h) h = nonlinearity(h) h = self.conv2(h) if self.in_dim != self.out_dim: if self.use_conv_shortcut: x = self.conv_shortcut(x) else: x = self.nin_shortcut(x) return mp_sum(x, h, t=0.3) class AttnBlock1D(nn.Module): def __init__(self, in_channels, num_heads=1): super().__init__() self.in_channels = in_channels self.num_heads = num_heads self.qkv = ops.Conv1d(in_channels, in_channels * 3, kernel_size=1, padding=0, bias=False) self.proj_out = ops.Conv1d(in_channels, in_channels, kernel_size=1, padding=0, bias=False) self.optimized_attention = vae_attention() def forward(self, x): h = x y = self.qkv(h) y = y.reshape(y.shape[0], -1, 3, y.shape[-1]) q, k, v = normalize(y, dim=1).unbind(2) h = self.optimized_attention(q, k, v) h = self.proj_out(h) return mp_sum(x, h, t=0.3) class Upsample1D(nn.Module): def __init__(self, in_channels, with_conv): super().__init__() self.with_conv = with_conv if self.with_conv: self.conv = ops.Conv1d(in_channels, in_channels, kernel_size=3, padding=1, bias=False) def forward(self, x): x = F.interpolate(x, scale_factor=2.0, mode='nearest-exact') # support 3D tensor(B,C,T) if self.with_conv: x = self.conv(x) return x class Downsample1D(nn.Module): def __init__(self, in_channels, with_conv): super().__init__() self.with_conv = with_conv if self.with_conv: # no asymmetric padding in torch conv, must do it ourselves self.conv1 = ops.Conv1d(in_channels, in_channels, kernel_size=1, padding=0, bias=False) self.conv2 = ops.Conv1d(in_channels, in_channels, kernel_size=1, padding=0, bias=False) def forward(self, x): if self.with_conv: x = self.conv1(x) x = F.avg_pool1d(x, kernel_size=2, stride=2) if self.with_conv: x = self.conv2(x) return x
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/ldm/mmaudio/vae/vae_modules.py", "license": "GNU General Public License v3.0", "lines": 90, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy_api_nodes/nodes_sora.py
from typing import Optional import torch from pydantic import BaseModel, Field from typing_extensions import override from comfy_api.latest import IO, ComfyExtension from comfy_api_nodes.util import ( ApiEndpoint, download_url_to_video_output, get_number_of_images, poll_op, sync_op, tensor_to_bytesio, ) class Sora2GenerationRequest(BaseModel): prompt: str = Field(...) model: str = Field(...) seconds: str = Field(...) size: str = Field(...) class Sora2GenerationResponse(BaseModel): id: str = Field(...) error: Optional[dict] = Field(None) status: Optional[str] = Field(None) class OpenAIVideoSora2(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="OpenAIVideoSora2", display_name="OpenAI Sora - Video", category="api node/video/Sora", description="OpenAI video and audio generation.", inputs=[ IO.Combo.Input( "model", options=["sora-2", "sora-2-pro"], default="sora-2", ), IO.String.Input( "prompt", multiline=True, default="", tooltip="Guiding text; may be empty if an input image is present.", ), IO.Combo.Input( "size", options=[ "720x1280", "1280x720", "1024x1792", "1792x1024", ], default="1280x720", ), IO.Combo.Input( "duration", options=[4, 8, 12], default=8, ), IO.Image.Input( "image", optional=True, ), IO.Int.Input( "seed", default=0, min=0, max=2147483647, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, optional=True, tooltip="Seed to determine if node should re-run; " "actual results are nondeterministic regardless of seed.", ), ], outputs=[ IO.Video.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["model", "size", "duration"]), expr=""" ( $m := widgets.model; $size := widgets.size; $dur := widgets.duration; $isPro := $contains($m, "sora-2-pro"); $isSora2 := $contains($m, "sora-2"); $isProSize := ($size = "1024x1792" or $size = "1792x1024"); $perSec := $isPro ? ($isProSize ? 0.5 : 0.3) : $isSora2 ? 0.1 : ($isProSize ? 0.5 : 0.1); {"type":"usd","usd": $round($perSec * $dur, 2)} ) """, ), ) @classmethod async def execute( cls, model: str, prompt: str, size: str = "1280x720", duration: int = 8, seed: int = 0, image: Optional[torch.Tensor] = None, ): if model == "sora-2" and size not in ("720x1280", "1280x720"): raise ValueError("Invalid size for sora-2 model, only 720x1280 and 1280x720 are supported.") files_input = None if image is not None: if get_number_of_images(image) != 1: raise ValueError("Currently only one input image is supported.") files_input = {"input_reference": ("image.png", tensor_to_bytesio(image), "image/png")} initial_response = await sync_op( cls, endpoint=ApiEndpoint(path="/proxy/openai/v1/videos", method="POST"), data=Sora2GenerationRequest( model=model, prompt=prompt, seconds=str(duration), size=size, ), files=files_input, response_model=Sora2GenerationResponse, content_type="multipart/form-data", ) if initial_response.error: raise Exception(initial_response.error["message"]) model_time_multiplier = 1 if model == "sora-2" else 2 await poll_op( cls, poll_endpoint=ApiEndpoint(path=f"/proxy/openai/v1/videos/{initial_response.id}"), response_model=Sora2GenerationResponse, status_extractor=lambda x: x.status, poll_interval=8.0, estimated_duration=int(45 * (duration / 4) * model_time_multiplier), ) return IO.NodeOutput( await download_url_to_video_output(f"/proxy/openai/v1/videos/{initial_response.id}/content", cls=cls), ) class OpenAISoraExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[IO.ComfyNode]]: return [ OpenAIVideoSora2, ] async def comfy_entrypoint() -> OpenAISoraExtension: return OpenAISoraExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api_nodes/nodes_sora.py", "license": "GNU General Public License v3.0", "lines": 154, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy_extras/nodes_eps.py
import torch from typing_extensions import override from comfy.k_diffusion.sampling import sigma_to_half_log_snr from comfy_api.latest import ComfyExtension, io class EpsilonScaling(io.ComfyNode): """ Implements the Epsilon Scaling method from 'Elucidating the Exposure Bias in Diffusion Models' (https://arxiv.org/abs/2308.15321v6). This method mitigates exposure bias by scaling the predicted noise during sampling, which can significantly improve sample quality. This implementation uses the "uniform schedule" recommended by the paper for its practicality and effectiveness. """ @classmethod def define_schema(cls): return io.Schema( node_id="Epsilon Scaling", category="model_patches/unet", inputs=[ io.Model.Input("model"), io.Float.Input( "scaling_factor", default=1.005, min=0.5, max=1.5, step=0.001, display_mode=io.NumberDisplay.number, advanced=True, ), ], outputs=[ io.Model.Output(), ], ) @classmethod def execute(cls, model, scaling_factor) -> io.NodeOutput: # Prevent division by zero, though the UI's min value should prevent this. if scaling_factor == 0: scaling_factor = 1e-9 def epsilon_scaling_function(args): """ This function is applied after the CFG guidance has been calculated. It recalculates the denoised latent by scaling the predicted noise. """ denoised = args["denoised"] x = args["input"] noise_pred = x - denoised scaled_noise_pred = noise_pred / scaling_factor new_denoised = x - scaled_noise_pred return new_denoised # Clone the model patcher to avoid modifying the original model in place model_clone = model.clone() model_clone.set_model_sampler_post_cfg_function(epsilon_scaling_function) return io.NodeOutput(model_clone) def compute_tsr_rescaling_factor( snr: torch.Tensor, tsr_k: float, tsr_variance: float ) -> torch.Tensor: """Compute the rescaling score ratio in Temporal Score Rescaling. See equation (6) in https://arxiv.org/pdf/2510.01184v1. """ posinf_mask = torch.isposinf(snr) rescaling_factor = (snr * tsr_variance + 1) / (snr * tsr_variance / tsr_k + 1) return torch.where(posinf_mask, tsr_k, rescaling_factor) # when snr → inf, r = tsr_k class TemporalScoreRescaling(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="TemporalScoreRescaling", display_name="TSR - Temporal Score Rescaling", category="model_patches/unet", inputs=[ io.Model.Input("model"), io.Float.Input( "tsr_k", tooltip=( "Controls the rescaling strength.\n" "Lower k produces more detailed results; higher k produces smoother results in image generation. Setting k = 1 disables rescaling." ), default=0.95, min=0.01, max=100.0, step=0.001, display_mode=io.NumberDisplay.number, advanced=True, ), io.Float.Input( "tsr_sigma", tooltip=( "Controls how early rescaling takes effect.\n" "Larger values take effect earlier." ), default=1.0, min=0.01, max=100.0, step=0.001, display_mode=io.NumberDisplay.number, advanced=True, ), ], outputs=[ io.Model.Output( display_name="patched_model", ), ], description=( "[Post-CFG Function]\n" "TSR - Temporal Score Rescaling (2510.01184)\n\n" "Rescaling the model's score or noise to steer the sampling diversity.\n" ), ) @classmethod def execute(cls, model, tsr_k, tsr_sigma) -> io.NodeOutput: tsr_variance = tsr_sigma**2 def temporal_score_rescaling(args): denoised = args["denoised"] x = args["input"] sigma = args["sigma"] curr_model = args["model"] # No rescaling (r = 1) or no noise if tsr_k == 1 or sigma == 0: return denoised model_sampling = curr_model.current_patcher.get_model_object("model_sampling") half_log_snr = sigma_to_half_log_snr(sigma, model_sampling) snr = (2 * half_log_snr).exp() # No rescaling needed (r = 1) if snr == 0: return denoised rescaling_r = compute_tsr_rescaling_factor(snr, tsr_k, tsr_variance) # Derived from scaled_denoised = (x - r * sigma * noise) / alpha alpha = sigma * half_log_snr.exp() return torch.lerp(x / alpha, denoised, rescaling_r) m = model.clone() m.set_model_sampler_post_cfg_function(temporal_score_rescaling) return io.NodeOutput(m) class EpsilonScalingExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[io.ComfyNode]]: return [ EpsilonScaling, TemporalScoreRescaling, ] async def comfy_entrypoint() -> EpsilonScalingExtension: return EpsilonScalingExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_extras/nodes_eps.py", "license": "GNU General Public License v3.0", "lines": 142, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy_api_nodes/nodes_wan.py
import re from pydantic import BaseModel, Field from typing_extensions import override from comfy_api.latest import IO, ComfyExtension, Input from comfy_api_nodes.util import ( ApiEndpoint, audio_to_base64_string, download_url_to_image_tensor, download_url_to_video_output, get_number_of_images, poll_op, sync_op, tensor_to_base64_string, upload_video_to_comfyapi, validate_audio_duration, validate_video_duration, ) class Text2ImageInputField(BaseModel): prompt: str = Field(...) negative_prompt: str | None = Field(None) class Image2ImageInputField(BaseModel): prompt: str = Field(...) negative_prompt: str | None = Field(None) images: list[str] = Field(..., min_length=1, max_length=2) class Text2VideoInputField(BaseModel): prompt: str = Field(...) negative_prompt: str | None = Field(None) audio_url: str | None = Field(None) class Image2VideoInputField(BaseModel): prompt: str = Field(...) negative_prompt: str | None = Field(None) img_url: str = Field(...) audio_url: str | None = Field(None) class Reference2VideoInputField(BaseModel): prompt: str = Field(...) negative_prompt: str | None = Field(None) reference_video_urls: list[str] = Field(...) class Txt2ImageParametersField(BaseModel): size: str = Field(...) n: int = Field(1, description="Number of images to generate.") # we support only value=1 seed: int = Field(..., ge=0, le=2147483647) prompt_extend: bool = Field(True) watermark: bool = Field(False) class Image2ImageParametersField(BaseModel): size: str | None = Field(None) n: int = Field(1, description="Number of images to generate.") # we support only value=1 seed: int = Field(..., ge=0, le=2147483647) watermark: bool = Field(False) class Text2VideoParametersField(BaseModel): size: str = Field(...) seed: int = Field(..., ge=0, le=2147483647) duration: int = Field(5, ge=5, le=15) prompt_extend: bool = Field(True) watermark: bool = Field(False) audio: bool = Field(False, description="Whether to generate audio automatically.") shot_type: str = Field("single") class Image2VideoParametersField(BaseModel): resolution: str = Field(...) seed: int = Field(..., ge=0, le=2147483647) duration: int = Field(5, ge=5, le=15) prompt_extend: bool = Field(True) watermark: bool = Field(False) audio: bool = Field(False, description="Whether to generate audio automatically.") shot_type: str = Field("single") class Reference2VideoParametersField(BaseModel): size: str = Field(...) duration: int = Field(5, ge=5, le=15) shot_type: str = Field("single") seed: int = Field(..., ge=0, le=2147483647) watermark: bool = Field(False) class Text2ImageTaskCreationRequest(BaseModel): model: str = Field(...) input: Text2ImageInputField = Field(...) parameters: Txt2ImageParametersField = Field(...) class Image2ImageTaskCreationRequest(BaseModel): model: str = Field(...) input: Image2ImageInputField = Field(...) parameters: Image2ImageParametersField = Field(...) class Text2VideoTaskCreationRequest(BaseModel): model: str = Field(...) input: Text2VideoInputField = Field(...) parameters: Text2VideoParametersField = Field(...) class Image2VideoTaskCreationRequest(BaseModel): model: str = Field(...) input: Image2VideoInputField = Field(...) parameters: Image2VideoParametersField = Field(...) class Reference2VideoTaskCreationRequest(BaseModel): model: str = Field(...) input: Reference2VideoInputField = Field(...) parameters: Reference2VideoParametersField = Field(...) class TaskCreationOutputField(BaseModel): task_id: str = Field(...) task_status: str = Field(...) class TaskCreationResponse(BaseModel): output: TaskCreationOutputField | None = Field(None) request_id: str = Field(...) code: str | None = Field(None, description="Error code for the failed request.") message: str | None = Field(None, description="Details about the failed request.") class TaskResult(BaseModel): url: str | None = Field(None) code: str | None = Field(None) message: str | None = Field(None) class ImageTaskStatusOutputField(TaskCreationOutputField): task_id: str = Field(...) task_status: str = Field(...) results: list[TaskResult] | None = Field(None) class VideoTaskStatusOutputField(TaskCreationOutputField): task_id: str = Field(...) task_status: str = Field(...) video_url: str | None = Field(None) code: str | None = Field(None) message: str | None = Field(None) class ImageTaskStatusResponse(BaseModel): output: ImageTaskStatusOutputField | None = Field(None) request_id: str = Field(...) class VideoTaskStatusResponse(BaseModel): output: VideoTaskStatusOutputField | None = Field(None) request_id: str = Field(...) RES_IN_PARENS = re.compile(r"\((\d+)\s*[x×]\s*(\d+)\)") class WanTextToImageApi(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="WanTextToImageApi", display_name="Wan Text to Image", category="api node/image/Wan", description="Generates an image based on a text prompt.", inputs=[ IO.Combo.Input( "model", options=["wan2.5-t2i-preview"], default="wan2.5-t2i-preview", tooltip="Model to use.", ), IO.String.Input( "prompt", multiline=True, default="", tooltip="Prompt describing the elements and visual features. Supports English and Chinese.", ), IO.String.Input( "negative_prompt", multiline=True, default="", tooltip="Negative prompt describing what to avoid.", optional=True, ), IO.Int.Input( "width", default=1024, min=768, max=1440, step=32, optional=True, ), IO.Int.Input( "height", default=1024, min=768, max=1440, step=32, optional=True, ), IO.Int.Input( "seed", default=0, min=0, max=2147483647, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, tooltip="Seed to use for generation.", optional=True, ), IO.Boolean.Input( "prompt_extend", default=True, tooltip="Whether to enhance the prompt with AI assistance.", optional=True, advanced=True, ), IO.Boolean.Input( "watermark", default=False, tooltip="Whether to add an AI-generated watermark to the result.", optional=True, advanced=True, ), ], outputs=[ IO.Image.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.03}""", ), ) @classmethod async def execute( cls, model: str, prompt: str, negative_prompt: str = "", width: int = 1024, height: int = 1024, seed: int = 0, prompt_extend: bool = True, watermark: bool = False, ): initial_response = await sync_op( cls, ApiEndpoint(path="/proxy/wan/api/v1/services/aigc/text2image/image-synthesis", method="POST"), response_model=TaskCreationResponse, data=Text2ImageTaskCreationRequest( model=model, input=Text2ImageInputField(prompt=prompt, negative_prompt=negative_prompt), parameters=Txt2ImageParametersField( size=f"{width}*{height}", seed=seed, prompt_extend=prompt_extend, watermark=watermark, ), ), ) if not initial_response.output: raise Exception(f"An unknown error occurred: {initial_response.code} - {initial_response.message}") response = await poll_op( cls, ApiEndpoint(path=f"/proxy/wan/api/v1/tasks/{initial_response.output.task_id}"), response_model=ImageTaskStatusResponse, status_extractor=lambda x: x.output.task_status, estimated_duration=9, poll_interval=3, ) return IO.NodeOutput(await download_url_to_image_tensor(str(response.output.results[0].url))) class WanImageToImageApi(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="WanImageToImageApi", display_name="Wan Image to Image", category="api node/image/Wan", description="Generates an image from one or two input images and a text prompt. " "The output image is currently fixed at 1.6 MP, and its aspect ratio matches the input image(s).", inputs=[ IO.Combo.Input( "model", options=["wan2.5-i2i-preview"], default="wan2.5-i2i-preview", tooltip="Model to use.", ), IO.Image.Input( "image", tooltip="Single-image editing or multi-image fusion. Maximum 2 images.", ), IO.String.Input( "prompt", multiline=True, default="", tooltip="Prompt describing the elements and visual features. Supports English and Chinese.", ), IO.String.Input( "negative_prompt", multiline=True, default="", tooltip="Negative prompt describing what to avoid.", optional=True, ), # redo this later as an optional combo of recommended resolutions # IO.Int.Input( # "width", # default=1280, # min=384, # max=1440, # step=16, # optional=True, # ), # IO.Int.Input( # "height", # default=1280, # min=384, # max=1440, # step=16, # optional=True, # ), IO.Int.Input( "seed", default=0, min=0, max=2147483647, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, tooltip="Seed to use for generation.", optional=True, ), IO.Boolean.Input( "watermark", default=False, tooltip="Whether to add an AI-generated watermark to the result.", optional=True, advanced=True, ), ], outputs=[ IO.Image.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.03}""", ), ) @classmethod async def execute( cls, model: str, image: Input.Image, prompt: str, negative_prompt: str = "", # width: int = 1024, # height: int = 1024, seed: int = 0, watermark: bool = False, ): n_images = get_number_of_images(image) if n_images not in (1, 2): raise ValueError(f"Expected 1 or 2 input images, but got {n_images}.") images = [] for i in image: images.append("data:image/png;base64," + tensor_to_base64_string(i, total_pixels=4096 * 4096)) initial_response = await sync_op( cls, ApiEndpoint(path="/proxy/wan/api/v1/services/aigc/image2image/image-synthesis", method="POST"), response_model=TaskCreationResponse, data=Image2ImageTaskCreationRequest( model=model, input=Image2ImageInputField(prompt=prompt, negative_prompt=negative_prompt, images=images), parameters=Image2ImageParametersField( # size=f"{width}*{height}", seed=seed, watermark=watermark, ), ), ) if not initial_response.output: raise Exception(f"An unknown error occurred: {initial_response.code} - {initial_response.message}") response = await poll_op( cls, ApiEndpoint(path=f"/proxy/wan/api/v1/tasks/{initial_response.output.task_id}"), response_model=ImageTaskStatusResponse, status_extractor=lambda x: x.output.task_status, estimated_duration=42, poll_interval=4, ) return IO.NodeOutput(await download_url_to_image_tensor(str(response.output.results[0].url))) class WanTextToVideoApi(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="WanTextToVideoApi", display_name="Wan Text to Video", category="api node/video/Wan", description="Generates a video based on a text prompt.", inputs=[ IO.Combo.Input( "model", options=["wan2.5-t2v-preview", "wan2.6-t2v"], default="wan2.6-t2v", tooltip="Model to use.", ), IO.String.Input( "prompt", multiline=True, default="", tooltip="Prompt describing the elements and visual features. Supports English and Chinese.", ), IO.String.Input( "negative_prompt", multiline=True, default="", tooltip="Negative prompt describing what to avoid.", optional=True, ), IO.Combo.Input( "size", options=[ "480p: 1:1 (624x624)", "480p: 16:9 (832x480)", "480p: 9:16 (480x832)", "720p: 1:1 (960x960)", "720p: 16:9 (1280x720)", "720p: 9:16 (720x1280)", "720p: 4:3 (1088x832)", "720p: 3:4 (832x1088)", "1080p: 1:1 (1440x1440)", "1080p: 16:9 (1920x1080)", "1080p: 9:16 (1080x1920)", "1080p: 4:3 (1632x1248)", "1080p: 3:4 (1248x1632)", ], default="720p: 1:1 (960x960)", optional=True, ), IO.Int.Input( "duration", default=5, min=5, max=15, step=5, display_mode=IO.NumberDisplay.number, tooltip="A 15-second duration is available only for the Wan 2.6 model.", optional=True, ), IO.Audio.Input( "audio", optional=True, tooltip="Audio must contain a clear, loud voice, without extraneous noise or background music.", ), IO.Int.Input( "seed", default=0, min=0, max=2147483647, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, tooltip="Seed to use for generation.", optional=True, ), IO.Boolean.Input( "generate_audio", default=False, optional=True, tooltip="If no audio input is provided, generate audio automatically.", advanced=True, ), IO.Boolean.Input( "prompt_extend", default=True, tooltip="Whether to enhance the prompt with AI assistance.", optional=True, advanced=True, ), IO.Boolean.Input( "watermark", default=False, tooltip="Whether to add an AI-generated watermark to the result.", optional=True, advanced=True, ), IO.Combo.Input( "shot_type", options=["single", "multi"], tooltip="Specifies the shot type for the generated video, that is, whether the video is a " "single continuous shot or multiple shots with cuts. " "This parameter takes effect only when prompt_extend is True.", optional=True, advanced=True, ), ], outputs=[ IO.Video.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["duration", "size"]), expr=""" ( $ppsTable := { "480p": 0.05, "720p": 0.1, "1080p": 0.15 }; $resKey := $substringBefore(widgets.size, ":"); $pps := $lookup($ppsTable, $resKey); { "type": "usd", "usd": $round($pps * widgets.duration, 2) } ) """, ), ) @classmethod async def execute( cls, model: str, prompt: str, negative_prompt: str = "", size: str = "720p: 1:1 (960x960)", duration: int = 5, audio: Input.Audio | None = None, seed: int = 0, generate_audio: bool = False, prompt_extend: bool = True, watermark: bool = False, shot_type: str = "single", ): if "480p" in size and model == "wan2.6-t2v": raise ValueError("The Wan 2.6 model does not support 480p.") if duration == 15 and model == "wan2.5-t2v-preview": raise ValueError("A 15-second duration is supported only by the Wan 2.6 model.") width, height = RES_IN_PARENS.search(size).groups() audio_url = None if audio is not None: validate_audio_duration(audio, 3.0, 29.0) audio_url = "data:audio/mp3;base64," + audio_to_base64_string(audio, "mp3", "libmp3lame") initial_response = await sync_op( cls, ApiEndpoint(path="/proxy/wan/api/v1/services/aigc/video-generation/video-synthesis", method="POST"), response_model=TaskCreationResponse, data=Text2VideoTaskCreationRequest( model=model, input=Text2VideoInputField(prompt=prompt, negative_prompt=negative_prompt, audio_url=audio_url), parameters=Text2VideoParametersField( size=f"{width}*{height}", duration=duration, seed=seed, audio=generate_audio, prompt_extend=prompt_extend, watermark=watermark, shot_type=shot_type, ), ), ) if not initial_response.output: raise Exception(f"An unknown error occurred: {initial_response.code} - {initial_response.message}") response = await poll_op( cls, ApiEndpoint(path=f"/proxy/wan/api/v1/tasks/{initial_response.output.task_id}"), response_model=VideoTaskStatusResponse, status_extractor=lambda x: x.output.task_status, estimated_duration=120 * int(duration / 5), poll_interval=6, ) return IO.NodeOutput(await download_url_to_video_output(response.output.video_url)) class WanImageToVideoApi(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="WanImageToVideoApi", display_name="Wan Image to Video", category="api node/video/Wan", description="Generates a video from the first frame and a text prompt.", inputs=[ IO.Combo.Input( "model", options=["wan2.5-i2v-preview", "wan2.6-i2v"], default="wan2.6-i2v", tooltip="Model to use.", ), IO.Image.Input( "image", ), IO.String.Input( "prompt", multiline=True, default="", tooltip="Prompt describing the elements and visual features. Supports English and Chinese.", ), IO.String.Input( "negative_prompt", multiline=True, default="", tooltip="Negative prompt describing what to avoid.", optional=True, ), IO.Combo.Input( "resolution", options=[ "480P", "720P", "1080P", ], default="720P", optional=True, ), IO.Int.Input( "duration", default=5, min=5, max=15, step=5, display_mode=IO.NumberDisplay.number, tooltip="Duration 15 available only for WAN2.6 model.", optional=True, ), IO.Audio.Input( "audio", optional=True, tooltip="Audio must contain a clear, loud voice, without extraneous noise or background music.", ), IO.Int.Input( "seed", default=0, min=0, max=2147483647, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, tooltip="Seed to use for generation.", optional=True, ), IO.Boolean.Input( "generate_audio", default=False, optional=True, tooltip="If no audio input is provided, generate audio automatically.", advanced=True, ), IO.Boolean.Input( "prompt_extend", default=True, tooltip="Whether to enhance the prompt with AI assistance.", optional=True, advanced=True, ), IO.Boolean.Input( "watermark", default=False, tooltip="Whether to add an AI-generated watermark to the result.", optional=True, advanced=True, ), IO.Combo.Input( "shot_type", options=["single", "multi"], tooltip="Specifies the shot type for the generated video, that is, whether the video is a " "single continuous shot or multiple shots with cuts. " "This parameter takes effect only when prompt_extend is True.", optional=True, advanced=True, ), ], outputs=[ IO.Video.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["duration", "resolution"]), expr=""" ( $ppsTable := { "480p": 0.05, "720p": 0.1, "1080p": 0.15 }; $pps := $lookup($ppsTable, widgets.resolution); { "type": "usd", "usd": $round($pps * widgets.duration, 2) } ) """, ), ) @classmethod async def execute( cls, model: str, image: Input.Image, prompt: str, negative_prompt: str = "", resolution: str = "720P", duration: int = 5, audio: Input.Audio | None = None, seed: int = 0, generate_audio: bool = False, prompt_extend: bool = True, watermark: bool = False, shot_type: str = "single", ): if get_number_of_images(image) != 1: raise ValueError("Exactly one input image is required.") if "480P" in resolution and model == "wan2.6-i2v": raise ValueError("The Wan 2.6 model does not support 480P.") if duration == 15 and model == "wan2.5-i2v-preview": raise ValueError("A 15-second duration is supported only by the Wan 2.6 model.") image_url = "data:image/png;base64," + tensor_to_base64_string(image, total_pixels=2000 * 2000) audio_url = None if audio is not None: validate_audio_duration(audio, 3.0, 29.0) audio_url = "data:audio/mp3;base64," + audio_to_base64_string(audio, "mp3", "libmp3lame") initial_response = await sync_op( cls, ApiEndpoint(path="/proxy/wan/api/v1/services/aigc/video-generation/video-synthesis", method="POST"), response_model=TaskCreationResponse, data=Image2VideoTaskCreationRequest( model=model, input=Image2VideoInputField( prompt=prompt, negative_prompt=negative_prompt, img_url=image_url, audio_url=audio_url ), parameters=Image2VideoParametersField( resolution=resolution, duration=duration, seed=seed, audio=generate_audio, prompt_extend=prompt_extend, watermark=watermark, shot_type=shot_type, ), ), ) if not initial_response.output: raise Exception(f"An unknown error occurred: {initial_response.code} - {initial_response.message}") response = await poll_op( cls, ApiEndpoint(path=f"/proxy/wan/api/v1/tasks/{initial_response.output.task_id}"), response_model=VideoTaskStatusResponse, status_extractor=lambda x: x.output.task_status, estimated_duration=120 * int(duration / 5), poll_interval=6, ) return IO.NodeOutput(await download_url_to_video_output(response.output.video_url)) class WanReferenceVideoApi(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="WanReferenceVideoApi", display_name="Wan Reference to Video", category="api node/video/Wan", description="Use the character and voice from input videos, combined with a prompt, " "to generate a new video that maintains character consistency.", inputs=[ IO.Combo.Input("model", options=["wan2.6-r2v"]), IO.String.Input( "prompt", multiline=True, default="", tooltip="Prompt describing the elements and visual features. Supports English and Chinese. " "Use identifiers such as `character1` and `character2` to refer to the reference characters.", ), IO.String.Input( "negative_prompt", multiline=True, default="", tooltip="Negative prompt describing what to avoid.", ), IO.Autogrow.Input( "reference_videos", template=IO.Autogrow.TemplateNames( IO.Video.Input("reference_video"), names=["character1", "character2", "character3"], min=1, ), ), IO.Combo.Input( "size", options=[ "720p: 1:1 (960x960)", "720p: 16:9 (1280x720)", "720p: 9:16 (720x1280)", "720p: 4:3 (1088x832)", "720p: 3:4 (832x1088)", "1080p: 1:1 (1440x1440)", "1080p: 16:9 (1920x1080)", "1080p: 9:16 (1080x1920)", "1080p: 4:3 (1632x1248)", "1080p: 3:4 (1248x1632)", ], ), IO.Int.Input( "duration", default=5, min=5, max=10, step=5, display_mode=IO.NumberDisplay.slider, ), IO.Int.Input( "seed", default=0, min=0, max=2147483647, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, ), IO.Combo.Input( "shot_type", options=["single", "multi"], tooltip="Specifies the shot type for the generated video, that is, whether the video is a " "single continuous shot or multiple shots with cuts.", advanced=True, ), IO.Boolean.Input( "watermark", default=False, tooltip="Whether to add an AI-generated watermark to the result.", advanced=True, ), ], outputs=[ IO.Video.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["size", "duration"]), expr=""" ( $rate := $contains(widgets.size, "1080p") ? 0.15 : 0.10; $inputMin := 2 * $rate; $inputMax := 5 * $rate; $outputPrice := widgets.duration * $rate; { "type": "range_usd", "min_usd": $inputMin + $outputPrice, "max_usd": $inputMax + $outputPrice } ) """, ), ) @classmethod async def execute( cls, model: str, prompt: str, negative_prompt: str, reference_videos: IO.Autogrow.Type, size: str, duration: int, seed: int, shot_type: str, watermark: bool, ): reference_video_urls = [] for i in reference_videos: validate_video_duration(reference_videos[i], min_duration=2, max_duration=30) for i in reference_videos: reference_video_urls.append(await upload_video_to_comfyapi(cls, reference_videos[i])) width, height = RES_IN_PARENS.search(size).groups() initial_response = await sync_op( cls, ApiEndpoint(path="/proxy/wan/api/v1/services/aigc/video-generation/video-synthesis", method="POST"), response_model=TaskCreationResponse, data=Reference2VideoTaskCreationRequest( model=model, input=Reference2VideoInputField( prompt=prompt, negative_prompt=negative_prompt, reference_video_urls=reference_video_urls ), parameters=Reference2VideoParametersField( size=f"{width}*{height}", duration=duration, shot_type=shot_type, watermark=watermark, seed=seed, ), ), ) if not initial_response.output: raise Exception(f"An unknown error occurred: {initial_response.code} - {initial_response.message}") response = await poll_op( cls, ApiEndpoint(path=f"/proxy/wan/api/v1/tasks/{initial_response.output.task_id}"), response_model=VideoTaskStatusResponse, status_extractor=lambda x: x.output.task_status, poll_interval=6, max_poll_attempts=280, ) return IO.NodeOutput(await download_url_to_video_output(response.output.video_url)) class WanApiExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[IO.ComfyNode]]: return [ WanTextToImageApi, WanImageToImageApi, WanTextToVideoApi, WanImageToVideoApi, WanReferenceVideoApi, ] async def comfy_entrypoint() -> WanApiExtension: return WanApiExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api_nodes/nodes_wan.py", "license": "GNU General Public License v3.0", "lines": 884, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy/ldm/wan/model_animate.py
from torch import nn import torch from typing import Tuple, Optional from einops import rearrange import torch.nn.functional as F import math from .model import WanModel, sinusoidal_embedding_1d from comfy.ldm.modules.attention import optimized_attention import comfy.model_management class CausalConv1d(nn.Module): def __init__(self, chan_in, chan_out, kernel_size=3, stride=1, dilation=1, pad_mode="replicate", operations=None, **kwargs): super().__init__() self.pad_mode = pad_mode padding = (kernel_size - 1, 0) # T self.time_causal_padding = padding self.conv = operations.Conv1d(chan_in, chan_out, kernel_size, stride=stride, dilation=dilation, **kwargs) def forward(self, x): x = F.pad(x, self.time_causal_padding, mode=self.pad_mode) return self.conv(x) class FaceEncoder(nn.Module): def __init__(self, in_dim: int, hidden_dim: int, num_heads=int, dtype=None, device=None, operations=None): factory_kwargs = {"dtype": dtype, "device": device} super().__init__() self.num_heads = num_heads self.conv1_local = CausalConv1d(in_dim, 1024 * num_heads, 3, stride=1, operations=operations, **factory_kwargs) self.norm1 = operations.LayerNorm(hidden_dim // 8, elementwise_affine=False, eps=1e-6, **factory_kwargs) self.act = nn.SiLU() self.conv2 = CausalConv1d(1024, 1024, 3, stride=2, operations=operations, **factory_kwargs) self.conv3 = CausalConv1d(1024, 1024, 3, stride=2, operations=operations, **factory_kwargs) self.out_proj = operations.Linear(1024, hidden_dim, **factory_kwargs) self.norm1 = operations.LayerNorm(1024, elementwise_affine=False, eps=1e-6, **factory_kwargs) self.norm2 = operations.LayerNorm(1024, elementwise_affine=False, eps=1e-6, **factory_kwargs) self.norm3 = operations.LayerNorm(1024, elementwise_affine=False, eps=1e-6, **factory_kwargs) self.padding_tokens = nn.Parameter(torch.empty(1, 1, 1, hidden_dim, **factory_kwargs)) def forward(self, x): x = rearrange(x, "b t c -> b c t") b, c, t = x.shape x = self.conv1_local(x) x = rearrange(x, "b (n c) t -> (b n) t c", n=self.num_heads) x = self.norm1(x) x = self.act(x) x = rearrange(x, "b t c -> b c t") x = self.conv2(x) x = rearrange(x, "b c t -> b t c") x = self.norm2(x) x = self.act(x) x = rearrange(x, "b t c -> b c t") x = self.conv3(x) x = rearrange(x, "b c t -> b t c") x = self.norm3(x) x = self.act(x) x = self.out_proj(x) x = rearrange(x, "(b n) t c -> b t n c", b=b) padding = comfy.model_management.cast_to(self.padding_tokens, dtype=x.dtype, device=x.device).repeat(b, x.shape[1], 1, 1) x = torch.cat([x, padding], dim=-2) x_local = x.clone() return x_local def get_norm_layer(norm_layer, operations=None): """ Get the normalization layer. Args: norm_layer (str): The type of normalization layer. Returns: norm_layer (nn.Module): The normalization layer. """ if norm_layer == "layer": return operations.LayerNorm elif norm_layer == "rms": return operations.RMSNorm else: raise NotImplementedError(f"Norm layer {norm_layer} is not implemented") class FaceAdapter(nn.Module): def __init__( self, hidden_dim: int, heads_num: int, qk_norm: bool = True, qk_norm_type: str = "rms", num_adapter_layers: int = 1, dtype=None, device=None, operations=None ): factory_kwargs = {"dtype": dtype, "device": device} super().__init__() self.hidden_size = hidden_dim self.heads_num = heads_num self.fuser_blocks = nn.ModuleList( [ FaceBlock( self.hidden_size, self.heads_num, qk_norm=qk_norm, qk_norm_type=qk_norm_type, operations=operations, **factory_kwargs, ) for _ in range(num_adapter_layers) ] ) def forward( self, x: torch.Tensor, motion_embed: torch.Tensor, idx: int, freqs_cis_q: Tuple[torch.Tensor, torch.Tensor] = None, freqs_cis_k: Tuple[torch.Tensor, torch.Tensor] = None, ) -> torch.Tensor: return self.fuser_blocks[idx](x, motion_embed, freqs_cis_q, freqs_cis_k) class FaceBlock(nn.Module): def __init__( self, hidden_size: int, heads_num: int, qk_norm: bool = True, qk_norm_type: str = "rms", qk_scale: float = None, dtype: Optional[torch.dtype] = None, device: Optional[torch.device] = None, operations=None ): factory_kwargs = {"device": device, "dtype": dtype} super().__init__() self.deterministic = False self.hidden_size = hidden_size self.heads_num = heads_num head_dim = hidden_size // heads_num self.scale = qk_scale or head_dim**-0.5 self.linear1_kv = operations.Linear(hidden_size, hidden_size * 2, **factory_kwargs) self.linear1_q = operations.Linear(hidden_size, hidden_size, **factory_kwargs) self.linear2 = operations.Linear(hidden_size, hidden_size, **factory_kwargs) qk_norm_layer = get_norm_layer(qk_norm_type, operations=operations) self.q_norm = ( qk_norm_layer(head_dim, elementwise_affine=True, eps=1e-6, **factory_kwargs) if qk_norm else nn.Identity() ) self.k_norm = ( qk_norm_layer(head_dim, elementwise_affine=True, eps=1e-6, **factory_kwargs) if qk_norm else nn.Identity() ) self.pre_norm_feat = operations.LayerNorm(hidden_size, elementwise_affine=False, eps=1e-6, **factory_kwargs) self.pre_norm_motion = operations.LayerNorm(hidden_size, elementwise_affine=False, eps=1e-6, **factory_kwargs) def forward( self, x: torch.Tensor, motion_vec: torch.Tensor, motion_mask: Optional[torch.Tensor] = None, # use_context_parallel=False, ) -> torch.Tensor: B, T, N, C = motion_vec.shape T_comp = T x_motion = self.pre_norm_motion(motion_vec) x_feat = self.pre_norm_feat(x) kv = self.linear1_kv(x_motion) q = self.linear1_q(x_feat) k, v = rearrange(kv, "B L N (K H D) -> K B L N H D", K=2, H=self.heads_num) q = rearrange(q, "B S (H D) -> B S H D", H=self.heads_num) # Apply QK-Norm if needed. q = self.q_norm(q).to(v) k = self.k_norm(k).to(v) k = rearrange(k, "B L N H D -> (B L) N H D") v = rearrange(v, "B L N H D -> (B L) N H D") q = rearrange(q, "B (L S) H D -> (B L) S (H D)", L=T_comp) attn = optimized_attention(q, k, v, heads=self.heads_num) attn = rearrange(attn, "(B L) S C -> B (L S) C", L=T_comp) output = self.linear2(attn) if motion_mask is not None: output = output * rearrange(motion_mask, "B T H W -> B (T H W)").unsqueeze(-1) return output # https://github.com/XPixelGroup/BasicSR/blob/8d56e3a045f9fb3e1d8872f92ee4a4f07f886b0a/basicsr/ops/upfirdn2d/upfirdn2d.py#L162 def upfirdn2d_native(input, kernel, up_x, up_y, down_x, down_y, pad_x0, pad_x1, pad_y0, pad_y1): _, minor, in_h, in_w = input.shape kernel_h, kernel_w = kernel.shape out = input.view(-1, minor, in_h, 1, in_w, 1) out = F.pad(out, [0, up_x - 1, 0, 0, 0, up_y - 1, 0, 0]) out = out.view(-1, minor, in_h * up_y, in_w * up_x) out = F.pad(out, [max(pad_x0, 0), max(pad_x1, 0), max(pad_y0, 0), max(pad_y1, 0)]) out = out[:, :, max(-pad_y0, 0): out.shape[2] - max(-pad_y1, 0), max(-pad_x0, 0): out.shape[3] - max(-pad_x1, 0)] out = out.reshape([-1, 1, in_h * up_y + pad_y0 + pad_y1, in_w * up_x + pad_x0 + pad_x1]) w = torch.flip(kernel, [0, 1]).view(1, 1, kernel_h, kernel_w) out = F.conv2d(out, w) out = out.reshape(-1, minor, in_h * up_y + pad_y0 + pad_y1 - kernel_h + 1, in_w * up_x + pad_x0 + pad_x1 - kernel_w + 1) return out[:, :, ::down_y, ::down_x] def upfirdn2d(input, kernel, up=1, down=1, pad=(0, 0)): return upfirdn2d_native(input, kernel, up, up, down, down, pad[0], pad[1], pad[0], pad[1]) # https://github.com/XPixelGroup/BasicSR/blob/8d56e3a045f9fb3e1d8872f92ee4a4f07f886b0a/basicsr/ops/fused_act/fused_act.py#L81 class FusedLeakyReLU(torch.nn.Module): def __init__(self, channel, negative_slope=0.2, scale=2 ** 0.5, dtype=None, device=None): super().__init__() self.bias = torch.nn.Parameter(torch.empty(1, channel, 1, 1, dtype=dtype, device=device)) self.negative_slope = negative_slope self.scale = scale def forward(self, input): return fused_leaky_relu(input, comfy.model_management.cast_to(self.bias, device=input.device, dtype=input.dtype), self.negative_slope, self.scale) def fused_leaky_relu(input, bias, negative_slope=0.2, scale=2 ** 0.5): return F.leaky_relu(input + bias, negative_slope) * scale class Blur(torch.nn.Module): def __init__(self, kernel, pad, dtype=None, device=None): super().__init__() kernel = torch.tensor(kernel, dtype=dtype, device=device) kernel = kernel[None, :] * kernel[:, None] kernel = kernel / kernel.sum() self.register_buffer('kernel', kernel) self.pad = pad def forward(self, input): return upfirdn2d(input, comfy.model_management.cast_to(self.kernel, dtype=input.dtype, device=input.device), pad=self.pad) #https://github.com/XPixelGroup/BasicSR/blob/8d56e3a045f9fb3e1d8872f92ee4a4f07f886b0a/basicsr/archs/stylegan2_arch.py#L590 class ScaledLeakyReLU(torch.nn.Module): def __init__(self, negative_slope=0.2): super().__init__() self.negative_slope = negative_slope def forward(self, input): return F.leaky_relu(input, negative_slope=self.negative_slope) # https://github.com/XPixelGroup/BasicSR/blob/8d56e3a045f9fb3e1d8872f92ee4a4f07f886b0a/basicsr/archs/stylegan2_arch.py#L605 class EqualConv2d(torch.nn.Module): def __init__(self, in_channel, out_channel, kernel_size, stride=1, padding=0, bias=True, dtype=None, device=None, operations=None): super().__init__() self.weight = torch.nn.Parameter(torch.empty(out_channel, in_channel, kernel_size, kernel_size, device=device, dtype=dtype)) self.scale = 1 / math.sqrt(in_channel * kernel_size ** 2) self.stride = stride self.padding = padding self.bias = torch.nn.Parameter(torch.empty(out_channel, device=device, dtype=dtype)) if bias else None def forward(self, input): if self.bias is None: bias = None else: bias = comfy.model_management.cast_to(self.bias, device=input.device, dtype=input.dtype) return F.conv2d(input, comfy.model_management.cast_to(self.weight, device=input.device, dtype=input.dtype) * self.scale, bias=bias, stride=self.stride, padding=self.padding) # https://github.com/XPixelGroup/BasicSR/blob/8d56e3a045f9fb3e1d8872f92ee4a4f07f886b0a/basicsr/archs/stylegan2_arch.py#L134 class EqualLinear(torch.nn.Module): def __init__(self, in_dim, out_dim, bias=True, bias_init=0, lr_mul=1, activation=None, dtype=None, device=None, operations=None): super().__init__() self.weight = torch.nn.Parameter(torch.empty(out_dim, in_dim, device=device, dtype=dtype)) self.bias = torch.nn.Parameter(torch.empty(out_dim, device=device, dtype=dtype)) if bias else None self.activation = activation self.scale = (1 / math.sqrt(in_dim)) * lr_mul self.lr_mul = lr_mul def forward(self, input): if self.bias is None: bias = None else: bias = comfy.model_management.cast_to(self.bias, device=input.device, dtype=input.dtype) * self.lr_mul if self.activation: out = F.linear(input, comfy.model_management.cast_to(self.weight, device=input.device, dtype=input.dtype) * self.scale) return fused_leaky_relu(out, bias) return F.linear(input, comfy.model_management.cast_to(self.weight, device=input.device, dtype=input.dtype) * self.scale, bias=bias) # https://github.com/XPixelGroup/BasicSR/blob/8d56e3a045f9fb3e1d8872f92ee4a4f07f886b0a/basicsr/archs/stylegan2_arch.py#L654 class ConvLayer(torch.nn.Sequential): def __init__(self, in_channel, out_channel, kernel_size, downsample=False, blur_kernel=[1, 3, 3, 1], bias=True, activate=True, dtype=None, device=None, operations=None): layers = [] if downsample: factor = 2 p = (len(blur_kernel) - factor) + (kernel_size - 1) layers.append(Blur(blur_kernel, pad=((p + 1) // 2, p // 2))) stride, padding = 2, 0 else: stride, padding = 1, kernel_size // 2 layers.append(EqualConv2d(in_channel, out_channel, kernel_size, padding=padding, stride=stride, bias=bias and not activate, dtype=dtype, device=device, operations=operations)) if activate: layers.append(FusedLeakyReLU(out_channel) if bias else ScaledLeakyReLU(0.2)) super().__init__(*layers) # https://github.com/XPixelGroup/BasicSR/blob/8d56e3a045f9fb3e1d8872f92ee4a4f07f886b0a/basicsr/archs/stylegan2_arch.py#L704 class ResBlock(torch.nn.Module): def __init__(self, in_channel, out_channel, dtype=None, device=None, operations=None): super().__init__() self.conv1 = ConvLayer(in_channel, in_channel, 3, dtype=dtype, device=device, operations=operations) self.conv2 = ConvLayer(in_channel, out_channel, 3, downsample=True, dtype=dtype, device=device, operations=operations) self.skip = ConvLayer(in_channel, out_channel, 1, downsample=True, activate=False, bias=False, dtype=dtype, device=device, operations=operations) def forward(self, input): out = self.conv2(self.conv1(input)) skip = self.skip(input) return (out + skip) / math.sqrt(2) class EncoderApp(torch.nn.Module): def __init__(self, w_dim=512, dtype=None, device=None, operations=None): super().__init__() kwargs = {"device": device, "dtype": dtype, "operations": operations} self.convs = torch.nn.ModuleList([ ConvLayer(3, 32, 1, **kwargs), ResBlock(32, 64, **kwargs), ResBlock(64, 128, **kwargs), ResBlock(128, 256, **kwargs), ResBlock(256, 512, **kwargs), ResBlock(512, 512, **kwargs), ResBlock(512, 512, **kwargs), ResBlock(512, 512, **kwargs), EqualConv2d(512, w_dim, 4, padding=0, bias=False, **kwargs) ]) def forward(self, x): h = x for conv in self.convs: h = conv(h) return h.squeeze(-1).squeeze(-1) class Encoder(torch.nn.Module): def __init__(self, dim=512, motion_dim=20, dtype=None, device=None, operations=None): super().__init__() self.net_app = EncoderApp(dim, dtype=dtype, device=device, operations=operations) self.fc = torch.nn.Sequential(*[EqualLinear(dim, dim, dtype=dtype, device=device, operations=operations) for _ in range(4)] + [EqualLinear(dim, motion_dim, dtype=dtype, device=device, operations=operations)]) def encode_motion(self, x): return self.fc(self.net_app(x)) class Direction(torch.nn.Module): def __init__(self, motion_dim, dtype=None, device=None, operations=None): super().__init__() self.weight = torch.nn.Parameter(torch.empty(512, motion_dim, device=device, dtype=dtype)) self.motion_dim = motion_dim def forward(self, input): stabilized_weight = comfy.model_management.cast_to(self.weight, device=input.device, dtype=input.dtype) + 1e-8 * torch.eye(512, self.motion_dim, device=input.device, dtype=input.dtype) Q, _ = torch.linalg.qr(stabilized_weight.float()) if input is None: return Q return torch.sum(input.unsqueeze(-1) * Q.T.to(input.dtype), dim=1) class Synthesis(torch.nn.Module): def __init__(self, motion_dim, dtype=None, device=None, operations=None): super().__init__() self.direction = Direction(motion_dim, dtype=dtype, device=device, operations=operations) class Generator(torch.nn.Module): def __init__(self, style_dim=512, motion_dim=20, dtype=None, device=None, operations=None): super().__init__() self.enc = Encoder(style_dim, motion_dim, dtype=dtype, device=device, operations=operations) self.dec = Synthesis(motion_dim, dtype=dtype, device=device, operations=operations) def get_motion(self, img): motion_feat = self.enc.encode_motion(img) return self.dec.direction(motion_feat) class AnimateWanModel(WanModel): r""" Wan diffusion backbone supporting both text-to-video and image-to-video. """ def __init__(self, model_type='animate', patch_size=(1, 2, 2), text_len=512, in_dim=16, dim=2048, ffn_dim=8192, freq_dim=256, text_dim=4096, out_dim=16, num_heads=16, num_layers=32, window_size=(-1, -1), qk_norm=True, cross_attn_norm=True, eps=1e-6, flf_pos_embed_token_number=None, motion_encoder_dim=512, image_model=None, device=None, dtype=None, operations=None, ): super().__init__(model_type='i2v', patch_size=patch_size, text_len=text_len, in_dim=in_dim, dim=dim, ffn_dim=ffn_dim, freq_dim=freq_dim, text_dim=text_dim, out_dim=out_dim, num_heads=num_heads, num_layers=num_layers, window_size=window_size, qk_norm=qk_norm, cross_attn_norm=cross_attn_norm, eps=eps, flf_pos_embed_token_number=flf_pos_embed_token_number, image_model=image_model, device=device, dtype=dtype, operations=operations) self.pose_patch_embedding = operations.Conv3d( 16, dim, kernel_size=patch_size, stride=patch_size, device=device, dtype=dtype ) self.motion_encoder = Generator(style_dim=512, motion_dim=20, device=device, dtype=dtype, operations=operations) self.face_adapter = FaceAdapter( heads_num=self.num_heads, hidden_dim=self.dim, num_adapter_layers=self.num_layers // 5, device=device, dtype=dtype, operations=operations ) self.face_encoder = FaceEncoder( in_dim=motion_encoder_dim, hidden_dim=self.dim, num_heads=4, device=device, dtype=dtype, operations=operations ) def after_patch_embedding(self, x, pose_latents, face_pixel_values): if pose_latents is not None: pose_latents = self.pose_patch_embedding(pose_latents) x[:, :, 1:pose_latents.shape[2] + 1] += pose_latents[:, :, :x.shape[2] - 1] if face_pixel_values is None: return x, None b, c, T, h, w = face_pixel_values.shape face_pixel_values = rearrange(face_pixel_values, "b c t h w -> (b t) c h w") encode_bs = 8 face_pixel_values_tmp = [] for i in range(math.ceil(face_pixel_values.shape[0] / encode_bs)): face_pixel_values_tmp.append(self.motion_encoder.get_motion(face_pixel_values[i * encode_bs: (i + 1) * encode_bs])) motion_vec = torch.cat(face_pixel_values_tmp) motion_vec = rearrange(motion_vec, "(b t) c -> b t c", t=T) motion_vec = self.face_encoder(motion_vec) B, L, H, C = motion_vec.shape pad_face = torch.zeros(B, 1, H, C).type_as(motion_vec) motion_vec = torch.cat([pad_face, motion_vec], dim=1) if motion_vec.shape[1] < x.shape[2]: B, L, H, C = motion_vec.shape pad = torch.zeros(B, x.shape[2] - motion_vec.shape[1], H, C).type_as(motion_vec) motion_vec = torch.cat([motion_vec, pad], dim=1) else: motion_vec = motion_vec[:, :x.shape[2]] return x, motion_vec def forward_orig( self, x, t, context, clip_fea=None, pose_latents=None, face_pixel_values=None, freqs=None, transformer_options={}, **kwargs, ): # embeddings x = self.patch_embedding(x.float()).to(x.dtype) x, motion_vec = self.after_patch_embedding(x, pose_latents, face_pixel_values) grid_sizes = x.shape[2:] x = x.flatten(2).transpose(1, 2) # time embeddings e = self.time_embedding( sinusoidal_embedding_1d(self.freq_dim, t.flatten()).to(dtype=x[0].dtype)) e = e.reshape(t.shape[0], -1, e.shape[-1]) e0 = self.time_projection(e).unflatten(2, (6, self.dim)) full_ref = None if self.ref_conv is not None: full_ref = kwargs.get("reference_latent", None) if full_ref is not None: full_ref = self.ref_conv(full_ref).flatten(2).transpose(1, 2) x = torch.concat((full_ref, x), dim=1) # context context = self.text_embedding(context) context_img_len = None if clip_fea is not None: if self.img_emb is not None: context_clip = self.img_emb(clip_fea) # bs x 257 x dim context = torch.concat([context_clip, context], dim=1) context_img_len = clip_fea.shape[-2] patches_replace = transformer_options.get("patches_replace", {}) blocks_replace = patches_replace.get("dit", {}) transformer_options["total_blocks"] = len(self.blocks) transformer_options["block_type"] = "double" for i, block in enumerate(self.blocks): transformer_options["block_index"] = i if ("double_block", i) in blocks_replace: def block_wrap(args): out = {} out["img"] = block(args["img"], context=args["txt"], e=args["vec"], freqs=args["pe"], context_img_len=context_img_len, transformer_options=args["transformer_options"]) return out out = blocks_replace[("double_block", i)]({"img": x, "txt": context, "vec": e0, "pe": freqs, "transformer_options": transformer_options}, {"original_block": block_wrap}) x = out["img"] else: x = block(x, e=e0, freqs=freqs, context=context, context_img_len=context_img_len, transformer_options=transformer_options) if i % 5 == 0 and motion_vec is not None: x = x + self.face_adapter.fuser_blocks[i // 5](x, motion_vec) # head x = self.head(x, e) if full_ref is not None: x = x[:, full_ref.shape[1]:] # unpatchify x = self.unpatchify(x, grid_sizes) return x
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/ldm/wan/model_animate.py", "license": "GNU General Public License v3.0", "lines": 443, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy/audio_encoders/whisper.py
import torch import torch.nn as nn import torch.nn.functional as F import torchaudio from typing import Optional from comfy.ldm.modules.attention import optimized_attention_masked import comfy.ops class WhisperFeatureExtractor(nn.Module): def __init__(self, n_mels=128, device=None): super().__init__() self.sample_rate = 16000 self.n_fft = 400 self.hop_length = 160 self.n_mels = n_mels self.chunk_length = 30 self.n_samples = 480000 self.mel_spectrogram = torchaudio.transforms.MelSpectrogram( sample_rate=self.sample_rate, n_fft=self.n_fft, hop_length=self.hop_length, n_mels=self.n_mels, f_min=0, f_max=8000, norm="slaney", mel_scale="slaney", ).to(device) def __call__(self, audio): audio = torch.mean(audio, dim=1) batch_size = audio.shape[0] processed_audio = [] for i in range(batch_size): aud = audio[i] if aud.shape[0] > self.n_samples: aud = aud[:self.n_samples] elif aud.shape[0] < self.n_samples: aud = F.pad(aud, (0, self.n_samples - aud.shape[0])) processed_audio.append(aud) audio = torch.stack(processed_audio) mel_spec = self.mel_spectrogram(audio.to(self.mel_spectrogram.spectrogram.window.device))[:, :, :-1].to(audio.device) log_mel_spec = torch.clamp(mel_spec, min=1e-10).log10() log_mel_spec = torch.maximum(log_mel_spec, log_mel_spec.max() - 8.0) log_mel_spec = (log_mel_spec + 4.0) / 4.0 return log_mel_spec class MultiHeadAttention(nn.Module): def __init__(self, d_model: int, n_heads: int, dtype=None, device=None, operations=None): super().__init__() assert d_model % n_heads == 0 self.d_model = d_model self.n_heads = n_heads self.d_k = d_model // n_heads self.q_proj = operations.Linear(d_model, d_model, dtype=dtype, device=device) self.k_proj = operations.Linear(d_model, d_model, bias=False, dtype=dtype, device=device) self.v_proj = operations.Linear(d_model, d_model, dtype=dtype, device=device) self.out_proj = operations.Linear(d_model, d_model, dtype=dtype, device=device) def forward( self, query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, mask: Optional[torch.Tensor] = None, ) -> torch.Tensor: batch_size, seq_len, _ = query.shape q = self.q_proj(query) k = self.k_proj(key) v = self.v_proj(value) attn_output = optimized_attention_masked(q, k, v, self.n_heads, mask) attn_output = self.out_proj(attn_output) return attn_output class EncoderLayer(nn.Module): def __init__(self, d_model: int, n_heads: int, d_ff: int, dtype=None, device=None, operations=None): super().__init__() self.self_attn = MultiHeadAttention(d_model, n_heads, dtype=dtype, device=device, operations=operations) self.self_attn_layer_norm = operations.LayerNorm(d_model, dtype=dtype, device=device) self.fc1 = operations.Linear(d_model, d_ff, dtype=dtype, device=device) self.fc2 = operations.Linear(d_ff, d_model, dtype=dtype, device=device) self.final_layer_norm = operations.LayerNorm(d_model, dtype=dtype, device=device) def forward( self, x: torch.Tensor, attention_mask: Optional[torch.Tensor] = None ) -> torch.Tensor: residual = x x = self.self_attn_layer_norm(x) x = self.self_attn(x, x, x, attention_mask) x = residual + x residual = x x = self.final_layer_norm(x) x = self.fc1(x) x = F.gelu(x) x = self.fc2(x) x = residual + x return x class AudioEncoder(nn.Module): def __init__( self, n_mels: int = 128, n_ctx: int = 1500, n_state: int = 1280, n_head: int = 20, n_layer: int = 32, dtype=None, device=None, operations=None ): super().__init__() self.conv1 = operations.Conv1d(n_mels, n_state, kernel_size=3, padding=1, dtype=dtype, device=device) self.conv2 = operations.Conv1d(n_state, n_state, kernel_size=3, stride=2, padding=1, dtype=dtype, device=device) self.embed_positions = operations.Embedding(n_ctx, n_state, dtype=dtype, device=device) self.layers = nn.ModuleList([ EncoderLayer(n_state, n_head, n_state * 4, dtype=dtype, device=device, operations=operations) for _ in range(n_layer) ]) self.layer_norm = operations.LayerNorm(n_state, dtype=dtype, device=device) def forward(self, x: torch.Tensor) -> torch.Tensor: x = F.gelu(self.conv1(x)) x = F.gelu(self.conv2(x)) x = x.transpose(1, 2) x = x + comfy.ops.cast_to_input(self.embed_positions.weight[:, :x.shape[1]], x) all_x = () for layer in self.layers: all_x += (x,) x = layer(x) x = self.layer_norm(x) all_x += (x,) return x, all_x class WhisperLargeV3(nn.Module): def __init__( self, n_mels: int = 128, n_audio_ctx: int = 1500, n_audio_state: int = 1280, n_audio_head: int = 20, n_audio_layer: int = 32, dtype=None, device=None, operations=None ): super().__init__() self.feature_extractor = WhisperFeatureExtractor(n_mels=n_mels, device=device) self.encoder = AudioEncoder( n_mels, n_audio_ctx, n_audio_state, n_audio_head, n_audio_layer, dtype=dtype, device=device, operations=operations ) def forward(self, audio): mel = self.feature_extractor(audio) x, all_x = self.encoder(mel) return x, all_x
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/audio_encoders/whisper.py", "license": "GNU General Public License v3.0", "lines": 147, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy/pixel_space_convert.py
import torch # "Fake" VAE that converts from IMAGE B, H, W, C and values on the scale of 0..1 # to LATENT B, C, H, W and values on the scale of -1..1. class PixelspaceConversionVAE(torch.nn.Module): def __init__(self): super().__init__() self.pixel_space_vae = torch.nn.Parameter(torch.tensor(1.0)) def encode(self, pixels: torch.Tensor, *_args, **_kwargs) -> torch.Tensor: return pixels def decode(self, samples: torch.Tensor, *_args, **_kwargs) -> torch.Tensor: return samples
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/pixel_space_convert.py", "license": "GNU General Public License v3.0", "lines": 11, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy/ldm/chroma_radiance/layers.py
# Adapted from https://github.com/lodestone-rock/flow from functools import lru_cache import torch from torch import nn class NerfEmbedder(nn.Module): """ An embedder module that combines input features with a 2D positional encoding that mimics the Discrete Cosine Transform (DCT). This module takes an input tensor of shape (B, P^2, C), where P is the patch size, and enriches it with positional information before projecting it to a new hidden size. """ def __init__( self, in_channels: int, hidden_size_input: int, max_freqs: int, dtype=None, device=None, operations=None, ): """ Initializes the NerfEmbedder. Args: in_channels (int): The number of channels in the input tensor. hidden_size_input (int): The desired dimension of the output embedding. max_freqs (int): The number of frequency components to use for both the x and y dimensions of the positional encoding. The total number of positional features will be max_freqs^2. """ super().__init__() self.dtype = dtype self.max_freqs = max_freqs self.hidden_size_input = hidden_size_input # A linear layer to project the concatenated input features and # positional encodings to the final output dimension. self.embedder = nn.Sequential( operations.Linear(in_channels + max_freqs**2, hidden_size_input, dtype=dtype, device=device) ) @lru_cache(maxsize=4) def fetch_pos(self, patch_size: int, device: torch.device, dtype: torch.dtype) -> torch.Tensor: """ Generates and caches 2D DCT-like positional embeddings for a given patch size. The LRU cache is a performance optimization that avoids recomputing the same positional grid on every forward pass. Args: patch_size (int): The side length of the square input patch. device: The torch device to create the tensors on. dtype: The torch dtype for the tensors. Returns: A tensor of shape (1, patch_size^2, max_freqs^2) containing the positional embeddings. """ # Create normalized 1D coordinate grids from 0 to 1. pos_x = torch.linspace(0, 1, patch_size, device=device, dtype=dtype) pos_y = torch.linspace(0, 1, patch_size, device=device, dtype=dtype) # Create a 2D meshgrid of coordinates. pos_y, pos_x = torch.meshgrid(pos_y, pos_x, indexing="ij") # Reshape positions to be broadcastable with frequencies. # Shape becomes (patch_size^2, 1, 1). pos_x = pos_x.reshape(-1, 1, 1) pos_y = pos_y.reshape(-1, 1, 1) # Create a 1D tensor of frequency values from 0 to max_freqs-1. freqs = torch.linspace(0, self.max_freqs - 1, self.max_freqs, dtype=dtype, device=device) # Reshape frequencies to be broadcastable for creating 2D basis functions. # freqs_x shape: (1, max_freqs, 1) # freqs_y shape: (1, 1, max_freqs) freqs_x = freqs[None, :, None] freqs_y = freqs[None, None, :] # A custom weighting coefficient, not part of standard DCT. # This seems to down-weight the contribution of higher-frequency interactions. coeffs = (1 + freqs_x * freqs_y) ** -1 # Calculate the 1D cosine basis functions for x and y coordinates. # This is the core of the DCT formulation. dct_x = torch.cos(pos_x * freqs_x * torch.pi) dct_y = torch.cos(pos_y * freqs_y * torch.pi) # Combine the 1D basis functions to create 2D basis functions by element-wise # multiplication, and apply the custom coefficients. Broadcasting handles the # combination of all (pos_x, freqs_x) with all (pos_y, freqs_y). # The result is flattened into a feature vector for each position. dct = (dct_x * dct_y * coeffs).view(1, -1, self.max_freqs ** 2) return dct def forward(self, inputs: torch.Tensor) -> torch.Tensor: """ Forward pass for the embedder. Args: inputs (Tensor): The input tensor of shape (B, P^2, C). Returns: Tensor: The output tensor of shape (B, P^2, hidden_size_input). """ # Get the batch size, number of pixels, and number of channels. B, P2, C = inputs.shape # Infer the patch side length from the number of pixels (P^2). patch_size = int(P2 ** 0.5) input_dtype = inputs.dtype inputs = inputs.to(dtype=self.dtype) # Fetch the pre-computed or cached positional embeddings. dct = self.fetch_pos(patch_size, inputs.device, self.dtype) # Repeat the positional embeddings for each item in the batch. dct = dct.repeat(B, 1, 1) # Concatenate the original input features with the positional embeddings # along the feature dimension. inputs = torch.cat((inputs, dct), dim=-1) # Project the combined tensor to the target hidden size. return self.embedder(inputs).to(dtype=input_dtype) class NerfGLUBlock(nn.Module): """ A NerfBlock using a Gated Linear Unit (GLU) like MLP. """ def __init__(self, hidden_size_s: int, hidden_size_x: int, mlp_ratio, dtype=None, device=None, operations=None): super().__init__() # The total number of parameters for the MLP is increased to accommodate # the gate, value, and output projection matrices. # We now need to generate parameters for 3 matrices. total_params = 3 * hidden_size_x**2 * mlp_ratio self.param_generator = operations.Linear(hidden_size_s, total_params, dtype=dtype, device=device) self.norm = operations.RMSNorm(hidden_size_x, dtype=dtype, device=device) self.mlp_ratio = mlp_ratio def forward(self, x: torch.Tensor, s: torch.Tensor) -> torch.Tensor: batch_size, num_x, hidden_size_x = x.shape mlp_params = self.param_generator(s) # Split the generated parameters into three parts for the gate, value, and output projection. fc1_gate_params, fc1_value_params, fc2_params = mlp_params.chunk(3, dim=-1) # Reshape the parameters into matrices for batch matrix multiplication. fc1_gate = fc1_gate_params.view(batch_size, hidden_size_x, hidden_size_x * self.mlp_ratio) fc1_value = fc1_value_params.view(batch_size, hidden_size_x, hidden_size_x * self.mlp_ratio) fc2 = fc2_params.view(batch_size, hidden_size_x * self.mlp_ratio, hidden_size_x) # Normalize the generated weight matrices as in the original implementation. fc1_gate = torch.nn.functional.normalize(fc1_gate, dim=-2) fc1_value = torch.nn.functional.normalize(fc1_value, dim=-2) fc2 = torch.nn.functional.normalize(fc2, dim=-2) res_x = x x = self.norm(x) # Apply the final output projection. x = torch.bmm(torch.nn.functional.silu(torch.bmm(x, fc1_gate)) * torch.bmm(x, fc1_value), fc2) return x + res_x class NerfFinalLayer(nn.Module): def __init__(self, hidden_size, out_channels, dtype=None, device=None, operations=None): super().__init__() self.norm = operations.RMSNorm(hidden_size, dtype=dtype, device=device) self.linear = operations.Linear(hidden_size, out_channels, dtype=dtype, device=device) def forward(self, x: torch.Tensor) -> torch.Tensor: # RMSNorm normalizes over the last dimension, but our channel dim (C) is at dim=1. # So we temporarily move the channel dimension to the end for the norm operation. return self.linear(self.norm(x.movedim(1, -1))).movedim(-1, 1) class NerfFinalLayerConv(nn.Module): def __init__(self, hidden_size: int, out_channels: int, dtype=None, device=None, operations=None): super().__init__() self.norm = operations.RMSNorm(hidden_size, dtype=dtype, device=device) self.conv = operations.Conv2d( in_channels=hidden_size, out_channels=out_channels, kernel_size=3, padding=1, dtype=dtype, device=device, ) def forward(self, x: torch.Tensor) -> torch.Tensor: # RMSNorm normalizes over the last dimension, but our channel dim (C) is at dim=1. # So we temporarily move the channel dimension to the end for the norm operation. return self.conv(self.norm(x.movedim(1, -1)).movedim(-1, 1))
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/ldm/chroma_radiance/layers.py", "license": "GNU General Public License v3.0", "lines": 161, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy/ldm/chroma_radiance/model.py
# Credits: # Original Flux code can be found on: https://github.com/black-forest-labs/flux # Chroma Radiance adaption referenced from https://github.com/lodestone-rock/flow from dataclasses import dataclass from typing import Optional import torch from torch import Tensor, nn from einops import repeat import comfy.ldm.common_dit from comfy.ldm.flux.layers import EmbedND, DoubleStreamBlock, SingleStreamBlock from comfy.ldm.chroma.model import Chroma, ChromaParams from comfy.ldm.chroma.layers import ( Approximator, ) from .layers import ( NerfEmbedder, NerfGLUBlock, NerfFinalLayer, NerfFinalLayerConv, ) @dataclass class ChromaRadianceParams(ChromaParams): patch_size: int nerf_hidden_size: int nerf_mlp_ratio: int nerf_depth: int nerf_max_freqs: int # Setting nerf_tile_size to 0 disables tiling. nerf_tile_size: int # Currently one of linear (legacy) or conv. nerf_final_head_type: str # None means use the same dtype as the model. nerf_embedder_dtype: Optional[torch.dtype] use_x0: bool class ChromaRadiance(Chroma): """ Transformer model for flow matching on sequences. """ def __init__(self, image_model=None, final_layer=True, dtype=None, device=None, operations=None, **kwargs): if operations is None: raise RuntimeError("Attempt to create ChromaRadiance object without setting operations") nn.Module.__init__(self) self.dtype = dtype params = ChromaRadianceParams(**kwargs) self.params = params self.patch_size = params.patch_size self.in_channels = params.in_channels self.out_channels = params.out_channels if params.hidden_size % params.num_heads != 0: raise ValueError( f"Hidden size {params.hidden_size} must be divisible by num_heads {params.num_heads}" ) pe_dim = params.hidden_size // params.num_heads if sum(params.axes_dim) != pe_dim: raise ValueError(f"Got {params.axes_dim} but expected positional dim {pe_dim}") self.hidden_size = params.hidden_size self.num_heads = params.num_heads self.in_dim = params.in_dim self.out_dim = params.out_dim self.hidden_dim = params.hidden_dim self.n_layers = params.n_layers self.pe_embedder = EmbedND(dim=pe_dim, theta=params.theta, axes_dim=params.axes_dim) self.img_in_patch = operations.Conv2d( params.in_channels, params.hidden_size, kernel_size=params.patch_size, stride=params.patch_size, bias=True, dtype=dtype, device=device, ) self.txt_in = operations.Linear(params.context_in_dim, self.hidden_size, dtype=dtype, device=device) # set as nn identity for now, will overwrite it later. self.distilled_guidance_layer = Approximator( in_dim=self.in_dim, hidden_dim=self.hidden_dim, out_dim=self.out_dim, n_layers=self.n_layers, dtype=dtype, device=device, operations=operations ) self.double_blocks = nn.ModuleList( [ DoubleStreamBlock( self.hidden_size, self.num_heads, mlp_ratio=params.mlp_ratio, qkv_bias=params.qkv_bias, modulation=False, dtype=dtype, device=device, operations=operations ) for _ in range(params.depth) ] ) self.single_blocks = nn.ModuleList( [ SingleStreamBlock( self.hidden_size, self.num_heads, mlp_ratio=params.mlp_ratio, modulation=False, dtype=dtype, device=device, operations=operations, ) for _ in range(params.depth_single_blocks) ] ) # pixel channel concat with DCT self.nerf_image_embedder = NerfEmbedder( in_channels=params.in_channels, hidden_size_input=params.nerf_hidden_size, max_freqs=params.nerf_max_freqs, dtype=params.nerf_embedder_dtype or dtype, device=device, operations=operations, ) self.nerf_blocks = nn.ModuleList([ NerfGLUBlock( hidden_size_s=params.hidden_size, hidden_size_x=params.nerf_hidden_size, mlp_ratio=params.nerf_mlp_ratio, dtype=dtype, device=device, operations=operations, ) for _ in range(params.nerf_depth) ]) if params.nerf_final_head_type == "linear": self.nerf_final_layer = NerfFinalLayer( params.nerf_hidden_size, out_channels=params.in_channels, dtype=dtype, device=device, operations=operations, ) elif params.nerf_final_head_type == "conv": self.nerf_final_layer_conv = NerfFinalLayerConv( params.nerf_hidden_size, out_channels=params.in_channels, dtype=dtype, device=device, operations=operations, ) else: errstr = f"Unsupported nerf_final_head_type {params.nerf_final_head_type}" raise ValueError(errstr) self.skip_mmdit = [] self.skip_dit = [] self.lite = False if params.use_x0: self.register_buffer("__x0__", torch.tensor([])) @property def _nerf_final_layer(self) -> nn.Module: if self.params.nerf_final_head_type == "linear": return self.nerf_final_layer if self.params.nerf_final_head_type == "conv": return self.nerf_final_layer_conv # Impossible to get here as we raise an error on unexpected types on initialization. raise NotImplementedError def img_in(self, img: Tensor) -> Tensor: img = self.img_in_patch(img) # -> [B, Hidden, H/P, W/P] # flatten into a sequence for the transformer. return img.flatten(2).transpose(1, 2) # -> [B, NumPatches, Hidden] def forward_nerf( self, img_orig: Tensor, img_out: Tensor, params: ChromaRadianceParams, ) -> Tensor: B, C, H, W = img_orig.shape num_patches = img_out.shape[1] patch_size = params.patch_size # Store the raw pixel values of each patch for the NeRF head later. # unfold creates patches: [B, C * P * P, NumPatches] nerf_pixels = nn.functional.unfold(img_orig, kernel_size=patch_size, stride=patch_size) nerf_pixels = nerf_pixels.transpose(1, 2) # -> [B, NumPatches, C * P * P] # Reshape for per-patch processing nerf_hidden = img_out.reshape(B * num_patches, params.hidden_size) nerf_pixels = nerf_pixels.reshape(B * num_patches, C, patch_size**2).transpose(1, 2) if params.nerf_tile_size > 0 and num_patches > params.nerf_tile_size: # Enable tiling if nerf_tile_size isn't 0 and we actually have more patches than # the tile size. img_dct = self.forward_tiled_nerf(nerf_hidden, nerf_pixels, B, C, num_patches, patch_size, params) else: # Get DCT-encoded pixel embeddings [pixel-dct] img_dct = self.nerf_image_embedder(nerf_pixels) # Pass through the dynamic MLP blocks (the NeRF) for block in self.nerf_blocks: img_dct = block(img_dct, nerf_hidden) # Reassemble the patches into the final image. img_dct = img_dct.transpose(1, 2) # -> [B*NumPatches, C, P*P] # Reshape to combine with batch dimension for fold img_dct = img_dct.reshape(B, num_patches, -1) # -> [B, NumPatches, C*P*P] img_dct = img_dct.transpose(1, 2) # -> [B, C*P*P, NumPatches] img_dct = nn.functional.fold( img_dct, output_size=(H, W), kernel_size=patch_size, stride=patch_size, ) return self._nerf_final_layer(img_dct) def forward_tiled_nerf( self, nerf_hidden: Tensor, nerf_pixels: Tensor, batch: int, channels: int, num_patches: int, patch_size: int, params: ChromaRadianceParams, ) -> Tensor: """ Processes the NeRF head in tiles to save memory. nerf_hidden has shape [B, L, D] nerf_pixels has shape [B, L, C * P * P] """ tile_size = params.nerf_tile_size output_tiles = [] # Iterate over the patches in tiles. The dimension L (num_patches) is at index 1. for i in range(0, num_patches, tile_size): end = min(i + tile_size, num_patches) # Slice the current tile from the input tensors nerf_hidden_tile = nerf_hidden[i * batch:end * batch] nerf_pixels_tile = nerf_pixels[i * batch:end * batch] # get DCT-encoded pixel embeddings [pixel-dct] img_dct_tile = self.nerf_image_embedder(nerf_pixels_tile) # pass through the dynamic MLP blocks (the NeRF) for block in self.nerf_blocks: img_dct_tile = block(img_dct_tile, nerf_hidden_tile) output_tiles.append(img_dct_tile) # Concatenate the processed tiles along the patch dimension return torch.cat(output_tiles, dim=0) def radiance_get_override_params(self, overrides: dict) -> ChromaRadianceParams: params = self.params if not overrides: return params params_dict = {k: getattr(params, k) for k in params.__dataclass_fields__} nullable_keys = frozenset(("nerf_embedder_dtype",)) bad_keys = tuple(k for k in overrides if k not in params_dict) if bad_keys: e = f"Unknown key(s) in transformer_options chroma_radiance_options: {', '.join(bad_keys)}" raise ValueError(e) bad_keys = tuple( k for k, v in overrides.items() if not isinstance(v, type(getattr(params, k))) and (v is not None or k not in nullable_keys) ) if bad_keys: e = f"Invalid value(s) in transformer_options chroma_radiance_options: {', '.join(bad_keys)}" raise ValueError(e) # At this point it's all valid keys and values so we can merge with the existing params. params_dict |= overrides return params.__class__(**params_dict) def _apply_x0_residual(self, predicted, noisy, timesteps): # non zero during training to prevent 0 div eps = 0.0 return (noisy - predicted) / (timesteps.view(-1,1,1,1) + eps) def _forward( self, x: Tensor, timestep: Tensor, context: Tensor, guidance: Optional[Tensor], control: Optional[dict]=None, transformer_options: dict={}, **kwargs: dict, ) -> Tensor: bs, c, h, w = x.shape img = comfy.ldm.common_dit.pad_to_patch_size(x, (self.patch_size, self.patch_size)) if img.ndim != 4: raise ValueError("Input img tensor must be in [B, C, H, W] format.") if context.ndim != 3: raise ValueError("Input txt tensors must have 3 dimensions.") params = self.radiance_get_override_params(transformer_options.get("chroma_radiance_options", {})) h_len = (img.shape[-2] // self.patch_size) w_len = (img.shape[-1] // self.patch_size) img_ids = torch.zeros((h_len, w_len, 3), device=x.device, dtype=x.dtype) img_ids[:, :, 1] = img_ids[:, :, 1] + torch.linspace(0, h_len - 1, steps=h_len, device=x.device, dtype=x.dtype).unsqueeze(1) img_ids[:, :, 2] = img_ids[:, :, 2] + torch.linspace(0, w_len - 1, steps=w_len, device=x.device, dtype=x.dtype).unsqueeze(0) img_ids = repeat(img_ids, "h w c -> b (h w) c", b=bs) txt_ids = torch.zeros((bs, context.shape[1], 3), device=x.device, dtype=x.dtype) img_out = self.forward_orig( img, img_ids, context, txt_ids, timestep, guidance, control, transformer_options, attn_mask=kwargs.get("attention_mask", None), ) out = self.forward_nerf(img, img_out, params)[:, :, :h, :w] # If x0 variant → v-pred, just return this instead if hasattr(self, "__x0__"): out = self._apply_x0_residual(out, img, timestep) return out
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/ldm/chroma_radiance/model.py", "license": "GNU General Public License v3.0", "lines": 294, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_extras/nodes_chroma_radiance.py
from typing_extensions import override from typing import Callable import torch import comfy.model_management from comfy_api.latest import ComfyExtension, io import nodes class EmptyChromaRadianceLatentImage(io.ComfyNode): @classmethod def define_schema(cls) -> io.Schema: return io.Schema( node_id="EmptyChromaRadianceLatentImage", category="latent/chroma_radiance", inputs=[ io.Int.Input(id="width", default=1024, min=16, max=nodes.MAX_RESOLUTION, step=16), io.Int.Input(id="height", default=1024, min=16, max=nodes.MAX_RESOLUTION, step=16), io.Int.Input(id="batch_size", default=1, min=1, max=4096), ], outputs=[io.Latent().Output()], ) @classmethod def execute(cls, *, width: int, height: int, batch_size: int=1) -> io.NodeOutput: latent = torch.zeros((batch_size, 3, height, width), device=comfy.model_management.intermediate_device()) return io.NodeOutput({"samples":latent}) class ChromaRadianceOptions(io.ComfyNode): @classmethod def define_schema(cls) -> io.Schema: return io.Schema( node_id="ChromaRadianceOptions", category="model_patches/chroma_radiance", description="Allows setting advanced options for the Chroma Radiance model.", inputs=[ io.Model.Input(id="model"), io.Boolean.Input( id="preserve_wrapper", default=True, tooltip="When enabled, will delegate to an existing model function wrapper if it exists. Generally should be left enabled.", ), io.Float.Input( id="start_sigma", default=1.0, min=0.0, max=1.0, tooltip="First sigma that these options will be in effect.", advanced=True, ), io.Float.Input( id="end_sigma", default=0.0, min=0.0, max=1.0, tooltip="Last sigma that these options will be in effect.", advanced=True, ), io.Int.Input( id="nerf_tile_size", default=-1, min=-1, tooltip="Allows overriding the default NeRF tile size. -1 means use the default (32). 0 means use non-tiling mode (may require a lot of VRAM).", advanced=True, ), ], outputs=[io.Model.Output()], ) @classmethod def execute( cls, *, model: io.Model.Type, preserve_wrapper: bool, start_sigma: float, end_sigma: float, nerf_tile_size: int, ) -> io.NodeOutput: radiance_options = {} if nerf_tile_size >= 0: radiance_options["nerf_tile_size"] = nerf_tile_size if not radiance_options: return io.NodeOutput(model) old_wrapper = model.model_options.get("model_function_wrapper") def model_function_wrapper(apply_model: Callable, args: dict) -> torch.Tensor: c = args["c"].copy() sigma = args["timestep"].max().detach().cpu().item() if end_sigma <= sigma <= start_sigma: transformer_options = c.get("transformer_options", {}).copy() transformer_options["chroma_radiance_options"] = radiance_options.copy() c["transformer_options"] = transformer_options if not (preserve_wrapper and old_wrapper): return apply_model(args["input"], args["timestep"], **c) return old_wrapper(apply_model, args | {"c": c}) model = model.clone() model.set_model_unet_function_wrapper(model_function_wrapper) return io.NodeOutput(model) class ChromaRadianceExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[io.ComfyNode]]: return [ EmptyChromaRadianceLatentImage, ChromaRadianceOptions, ] async def comfy_entrypoint() -> ChromaRadianceExtension: return ChromaRadianceExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_extras/nodes_chroma_radiance.py", "license": "GNU General Public License v3.0", "lines": 101, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy/ldm/hunyuan_video/vae_refiner.py
import torch import torch.nn as nn import torch.nn.functional as F from comfy.ldm.modules.diffusionmodules.model import ResnetBlock, AttnBlock, CarriedConv3d, Normalize, conv_carry_causal_3d, torch_cat_if_needed import comfy.ops import comfy.ldm.models.autoencoder import comfy.model_management ops = comfy.ops.disable_weight_init class RMS_norm(nn.Module): def __init__(self, dim): super().__init__() shape = (dim, 1, 1, 1) self.scale = dim**0.5 self.gamma = nn.Parameter(torch.empty(shape)) def forward(self, x): return F.normalize(x, dim=1) * self.scale * comfy.model_management.cast_to(self.gamma, dtype=x.dtype, device=x.device) class DnSmpl(nn.Module): def __init__(self, ic, oc, tds, refiner_vae, op): super().__init__() fct = 2 * 2 * 2 if tds else 1 * 2 * 2 assert oc % fct == 0 self.conv = op(ic, oc // fct, kernel_size=3, stride=1, padding=1) self.refiner_vae = refiner_vae self.tds = tds self.gs = fct * ic // oc def forward(self, x, conv_carry_in=None, conv_carry_out=None): r1 = 2 if self.tds else 1 h = conv_carry_causal_3d([x], self.conv, conv_carry_in, conv_carry_out) if self.tds and self.refiner_vae and conv_carry_in is None: hf = h[:, :, :1, :, :] b, c, f, ht, wd = hf.shape hf = hf.reshape(b, c, f, ht // 2, 2, wd // 2, 2) hf = hf.permute(0, 4, 6, 1, 2, 3, 5) hf = hf.reshape(b, 2 * 2 * c, f, ht // 2, wd // 2) hf = torch.cat([hf, hf], dim=1) h = h[:, :, 1:, :, :] xf = x[:, :, :1, :, :] b, ci, f, ht, wd = xf.shape xf = xf.reshape(b, ci, f, ht // 2, 2, wd // 2, 2) xf = xf.permute(0, 4, 6, 1, 2, 3, 5) xf = xf.reshape(b, 2 * 2 * ci, f, ht // 2, wd // 2) B, C, T, H, W = xf.shape xf = xf.view(B, hf.shape[1], self.gs // 2, T, H, W).mean(dim=2) x = x[:, :, 1:, :, :] if h.shape[2] == 0: return hf + xf b, c, frms, ht, wd = h.shape nf = frms // r1 h = h.reshape(b, c, nf, r1, ht // 2, 2, wd // 2, 2) h = h.permute(0, 3, 5, 7, 1, 2, 4, 6) h = h.reshape(b, r1 * 2 * 2 * c, nf, ht // 2, wd // 2) b, ci, frms, ht, wd = x.shape nf = frms // r1 x = x.reshape(b, ci, nf, r1, ht // 2, 2, wd // 2, 2) x = x.permute(0, 3, 5, 7, 1, 2, 4, 6) x = x.reshape(b, r1 * 2 * 2 * ci, nf, ht // 2, wd // 2) B, C, T, H, W = x.shape x = x.view(B, h.shape[1], self.gs, T, H, W).mean(dim=2) if self.tds and self.refiner_vae and conv_carry_in is None: h = torch.cat([hf, h], dim=2) x = torch.cat([xf, x], dim=2) return h + x class UpSmpl(nn.Module): def __init__(self, ic, oc, tus, refiner_vae, op): super().__init__() fct = 2 * 2 * 2 if tus else 1 * 2 * 2 self.conv = op(ic, oc * fct, kernel_size=3, stride=1, padding=1) self.refiner_vae = refiner_vae self.tus = tus self.rp = fct * oc // ic def forward(self, x, conv_carry_in=None, conv_carry_out=None): r1 = 2 if self.tus else 1 h = conv_carry_causal_3d([x], self.conv, conv_carry_in, conv_carry_out) if self.tus and self.refiner_vae and conv_carry_in is None: hf = h[:, :, :1, :, :] b, c, f, ht, wd = hf.shape nc = c // (2 * 2) hf = hf.reshape(b, 2, 2, nc, f, ht, wd) hf = hf.permute(0, 3, 4, 5, 1, 6, 2) hf = hf.reshape(b, nc, f, ht * 2, wd * 2) hf = hf[:, : hf.shape[1] // 2] h = h[:, :, 1:, :, :] xf = x[:, :, :1, :, :] b, ci, f, ht, wd = xf.shape xf = xf.repeat_interleave(repeats=self.rp // 2, dim=1) b, c, f, ht, wd = xf.shape nc = c // (2 * 2) xf = xf.reshape(b, 2, 2, nc, f, ht, wd) xf = xf.permute(0, 3, 4, 5, 1, 6, 2) xf = xf.reshape(b, nc, f, ht * 2, wd * 2) x = x[:, :, 1:, :, :] b, c, frms, ht, wd = h.shape nc = c // (r1 * 2 * 2) h = h.reshape(b, r1, 2, 2, nc, frms, ht, wd) h = h.permute(0, 4, 5, 1, 6, 2, 7, 3) h = h.reshape(b, nc, frms * r1, ht * 2, wd * 2) x = x.repeat_interleave(repeats=self.rp, dim=1) b, c, frms, ht, wd = x.shape nc = c // (r1 * 2 * 2) x = x.reshape(b, r1, 2, 2, nc, frms, ht, wd) x = x.permute(0, 4, 5, 1, 6, 2, 7, 3) x = x.reshape(b, nc, frms * r1, ht * 2, wd * 2) if self.tus and self.refiner_vae and conv_carry_in is None: h = torch.cat([hf, h], dim=2) x = torch.cat([xf, x], dim=2) return h + x class Encoder(nn.Module): def __init__(self, in_channels, z_channels, block_out_channels, num_res_blocks, ffactor_spatial, ffactor_temporal, downsample_match_channel=True, refiner_vae=True, **_): super().__init__() self.z_channels = z_channels self.block_out_channels = block_out_channels self.num_res_blocks = num_res_blocks self.ffactor_temporal = ffactor_temporal self.refiner_vae = refiner_vae if self.refiner_vae: conv_op = CarriedConv3d norm_op = RMS_norm else: conv_op = ops.Conv3d norm_op = Normalize self.conv_in = conv_op(in_channels, block_out_channels[0], 3, 1, 1) self.down = nn.ModuleList() ch = block_out_channels[0] depth = (ffactor_spatial >> 1).bit_length() depth_temporal = ((ffactor_spatial // self.ffactor_temporal) >> 1).bit_length() for i, tgt in enumerate(block_out_channels): stage = nn.Module() stage.block = nn.ModuleList([ResnetBlock(in_channels=ch if j == 0 else tgt, out_channels=tgt, temb_channels=0, conv_op=conv_op, norm_op=norm_op) for j in range(num_res_blocks)]) ch = tgt if i < depth: nxt = block_out_channels[i + 1] if i + 1 < len(block_out_channels) and downsample_match_channel else ch stage.downsample = DnSmpl(ch, nxt, tds=i >= depth_temporal, refiner_vae=self.refiner_vae, op=conv_op) ch = nxt self.down.append(stage) self.mid = nn.Module() self.mid.block_1 = ResnetBlock(in_channels=ch, out_channels=ch, conv_op=conv_op, norm_op=norm_op) self.mid.attn_1 = AttnBlock(ch, conv_op=ops.Conv3d, norm_op=norm_op) self.mid.block_2 = ResnetBlock(in_channels=ch, out_channels=ch, conv_op=conv_op, norm_op=norm_op) self.norm_out = norm_op(ch) self.conv_out = conv_op(ch, z_channels << 1, 3, 1, 1) self.regul = comfy.ldm.models.autoencoder.DiagonalGaussianRegularizer() def forward(self, x): if not self.refiner_vae and x.shape[2] == 1: x = x.expand(-1, -1, self.ffactor_temporal, -1, -1) if self.refiner_vae: xl = [x[:, :, :1, :, :]] if x.shape[2] > self.ffactor_temporal: xl += torch.split(x[:, :, 1: 1 + ((x.shape[2] - 1) // self.ffactor_temporal) * self.ffactor_temporal, :, :], self.ffactor_temporal * 2, dim=2) x = xl else: x = [x] out = [] conv_carry_in = None for i, x1 in enumerate(x): conv_carry_out = [] if i == len(x) - 1: conv_carry_out = None x1 = [ x1 ] x1 = conv_carry_causal_3d(x1, self.conv_in, conv_carry_in, conv_carry_out) for stage in self.down: for blk in stage.block: x1 = blk(x1, None, conv_carry_in, conv_carry_out) if hasattr(stage, 'downsample'): x1 = stage.downsample(x1, conv_carry_in, conv_carry_out) out.append(x1) conv_carry_in = conv_carry_out out = torch_cat_if_needed(out, dim=2) x = self.mid.block_2(self.mid.attn_1(self.mid.block_1(out))) del out b, c, t, h, w = x.shape grp = c // (self.z_channels << 1) skip = x.view(b, c // grp, grp, t, h, w).mean(2) out = conv_carry_causal_3d([F.silu(self.norm_out(x))], self.conv_out) + skip if self.refiner_vae: out = self.regul(out)[0] return out class Decoder(nn.Module): def __init__(self, z_channels, out_channels, block_out_channels, num_res_blocks, ffactor_spatial, ffactor_temporal, upsample_match_channel=True, refiner_vae=True, **_): super().__init__() block_out_channels = block_out_channels[::-1] self.z_channels = z_channels self.block_out_channels = block_out_channels self.num_res_blocks = num_res_blocks self.refiner_vae = refiner_vae if self.refiner_vae: conv_op = CarriedConv3d norm_op = RMS_norm else: conv_op = ops.Conv3d norm_op = Normalize ch = block_out_channels[0] self.conv_in = conv_op(z_channels, ch, kernel_size=3, stride=1, padding=1) self.mid = nn.Module() self.mid.block_1 = ResnetBlock(in_channels=ch, out_channels=ch, conv_op=conv_op, norm_op=norm_op) self.mid.attn_1 = AttnBlock(ch, conv_op=ops.Conv3d, norm_op=norm_op) self.mid.block_2 = ResnetBlock(in_channels=ch, out_channels=ch, conv_op=conv_op, norm_op=norm_op) self.up = nn.ModuleList() depth = (ffactor_spatial >> 1).bit_length() depth_temporal = (ffactor_temporal >> 1).bit_length() for i, tgt in enumerate(block_out_channels): stage = nn.Module() stage.block = nn.ModuleList([ResnetBlock(in_channels=ch if j == 0 else tgt, out_channels=tgt, temb_channels=0, conv_op=conv_op, norm_op=norm_op) for j in range(num_res_blocks + 1)]) ch = tgt if i < depth: nxt = block_out_channels[i + 1] if i + 1 < len(block_out_channels) and upsample_match_channel else ch stage.upsample = UpSmpl(ch, nxt, tus=i < depth_temporal, refiner_vae=self.refiner_vae, op=conv_op) ch = nxt self.up.append(stage) self.norm_out = norm_op(ch) self.conv_out = conv_op(ch, out_channels, 3, stride=1, padding=1) def forward(self, z): x = conv_carry_causal_3d([z], self.conv_in) + z.repeat_interleave(self.block_out_channels[0] // self.z_channels, 1) x = self.mid.block_2(self.mid.attn_1(self.mid.block_1(x))) if self.refiner_vae: x = torch.split(x, 2, dim=2) else: x = [ x ] out = [] conv_carry_in = None for i, x1 in enumerate(x): conv_carry_out = [] if i == len(x) - 1: conv_carry_out = None for stage in self.up: for blk in stage.block: x1 = blk(x1, None, conv_carry_in, conv_carry_out) if hasattr(stage, 'upsample'): x1 = stage.upsample(x1, conv_carry_in, conv_carry_out) x1 = [ F.silu(self.norm_out(x1)) ] x1 = conv_carry_causal_3d(x1, self.conv_out, conv_carry_in, conv_carry_out) out.append(x1) conv_carry_in = conv_carry_out del x out = torch_cat_if_needed(out, dim=2) if not self.refiner_vae: if z.shape[-3] == 1: out = out[:, :, -1:] return out
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/ldm/hunyuan_video/vae_refiner.py", "license": "GNU General Public License v3.0", "lines": 248, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy/text_encoders/hunyuan_image.py
from comfy import sd1_clip import comfy.text_encoders.llama from .qwen_image import QwenImageTokenizer, QwenImageTEModel from transformers import ByT5Tokenizer import os import re class ByT5SmallTokenizer(sd1_clip.SDTokenizer): def __init__(self, embedding_directory=None, tokenizer_data={}): tokenizer_path = os.path.join(os.path.dirname(os.path.realpath(__file__)), "byt5_tokenizer") super().__init__(tokenizer_path, pad_with_end=False, embedding_size=1472, embedding_key='byt5_small', tokenizer_class=ByT5Tokenizer, has_start_token=False, pad_to_max_length=False, max_length=99999999, min_length=1, tokenizer_data=tokenizer_data) class HunyuanImageTokenizer(QwenImageTokenizer): def __init__(self, embedding_directory=None, tokenizer_data={}): super().__init__(embedding_directory=embedding_directory, tokenizer_data=tokenizer_data) self.llama_template = "<|im_start|>system\nDescribe the image by detailing the color, shape, size, texture, quantity, text, spatial relationships of the objects and background:<|im_end|>\n<|im_start|>user\n{}<|im_end|>" # self.llama_template_images = "{}" self.byt5 = ByT5SmallTokenizer(embedding_directory=embedding_directory, tokenizer_data=tokenizer_data) def tokenize_with_weights(self, text:str, return_word_ids=False, **kwargs): out = super().tokenize_with_weights(text, return_word_ids, **kwargs) # ByT5 processing for HunyuanImage text_prompt_texts = [] pattern_quote_double = r'\"(.*?)\"' pattern_quote_chinese_single = r'‘(.*?)’' pattern_quote_chinese_double = r'“(.*?)”' matches_quote_double = re.findall(pattern_quote_double, text) matches_quote_chinese_single = re.findall(pattern_quote_chinese_single, text) matches_quote_chinese_double = re.findall(pattern_quote_chinese_double, text) text_prompt_texts.extend(matches_quote_double) text_prompt_texts.extend(matches_quote_chinese_single) text_prompt_texts.extend(matches_quote_chinese_double) if len(text_prompt_texts) > 0: out['byt5'] = self.byt5.tokenize_with_weights(''.join(map(lambda a: 'Text "{}". '.format(a), text_prompt_texts)), return_word_ids, **kwargs) return out class Qwen25_7BVLIModel(sd1_clip.SDClipModel): def __init__(self, device="cpu", layer="hidden", layer_idx=-3, dtype=None, attention_mask=True, model_options={}): llama_quantization_metadata = model_options.get("llama_quantization_metadata", None) if llama_quantization_metadata is not None: model_options = model_options.copy() model_options["quantization_metadata"] = llama_quantization_metadata super().__init__(device=device, layer=layer, layer_idx=layer_idx, textmodel_json_config={}, dtype=dtype, special_tokens={"pad": 151643}, layer_norm_hidden_state=False, model_class=comfy.text_encoders.llama.Qwen25_7BVLI, enable_attention_masks=attention_mask, return_attention_masks=attention_mask, model_options=model_options) class ByT5SmallModel(sd1_clip.SDClipModel): def __init__(self, device="cpu", layer="last", layer_idx=None, dtype=None, model_options={}): textmodel_json_config = os.path.join(os.path.dirname(os.path.realpath(__file__)), "byt5_config_small_glyph.json") super().__init__(device=device, layer=layer, layer_idx=layer_idx, textmodel_json_config=textmodel_json_config, dtype=dtype, model_options=model_options, special_tokens={"end": 1, "pad": 0}, model_class=comfy.text_encoders.t5.T5, enable_attention_masks=True, zero_out_masked=True) class HunyuanImageTEModel(QwenImageTEModel): def __init__(self, byt5=True, device="cpu", dtype=None, model_options={}): super(QwenImageTEModel, self).__init__(device=device, dtype=dtype, name="qwen25_7b", clip_model=Qwen25_7BVLIModel, model_options=model_options) if byt5: self.byt5_small = ByT5SmallModel(device=device, dtype=dtype, model_options=model_options) else: self.byt5_small = None def encode_token_weights(self, token_weight_pairs): tok_pairs = token_weight_pairs["qwen25_7b"][0] template_end = -1 if tok_pairs[0][0] == 27: if len(tok_pairs) > 36: # refiner prompt uses a fixed 36 template_end template_end = 36 cond, p, extra = super().encode_token_weights(token_weight_pairs, template_end=template_end) if self.byt5_small is not None and "byt5" in token_weight_pairs: out = self.byt5_small.encode_token_weights(token_weight_pairs["byt5"]) extra["conditioning_byt5small"] = out[0] return cond, p, extra def set_clip_options(self, options): super().set_clip_options(options) if self.byt5_small is not None: self.byt5_small.set_clip_options(options) def reset_clip_options(self): super().reset_clip_options() if self.byt5_small is not None: self.byt5_small.reset_clip_options() def load_sd(self, sd): if "encoder.block.0.layer.0.SelfAttention.o.weight" in sd: return self.byt5_small.load_sd(sd) else: return super().load_sd(sd) def te(byt5=True, dtype_llama=None, llama_quantization_metadata=None): class QwenImageTEModel_(HunyuanImageTEModel): def __init__(self, device="cpu", dtype=None, model_options={}): if llama_quantization_metadata is not None: model_options = model_options.copy() model_options["llama_quantization_metadata"] = llama_quantization_metadata if dtype_llama is not None: dtype = dtype_llama super().__init__(byt5=byt5, device=device, dtype=dtype, model_options=model_options) return QwenImageTEModel_
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/text_encoders/hunyuan_image.py", "license": "GNU General Public License v3.0", "lines": 84, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:tests-unit/server_test/test_cache_control.py
"""Tests for server cache control middleware""" import pytest from aiohttp import web from aiohttp.test_utils import make_mocked_request from typing import Dict, Any from middleware.cache_middleware import cache_control, ONE_HOUR, ONE_DAY, IMG_EXTENSIONS pytestmark = pytest.mark.asyncio # Apply asyncio mark to all tests # Test configuration data CACHE_SCENARIOS = [ # Image file scenarios { "name": "image_200_status", "path": "/test.jpg", "status": 200, "expected_cache": f"public, max-age={ONE_DAY}", "should_have_header": True, }, { "name": "image_404_status", "path": "/missing.jpg", "status": 404, "expected_cache": f"public, max-age={ONE_HOUR}", "should_have_header": True, }, # JavaScript/CSS scenarios { "name": "js_no_cache", "path": "/script.js", "status": 200, "expected_cache": "no-cache", "should_have_header": True, }, { "name": "css_no_cache", "path": "/styles.css", "status": 200, "expected_cache": "no-cache", "should_have_header": True, }, { "name": "index_json_no_cache", "path": "/api/index.json", "status": 200, "expected_cache": "no-cache", "should_have_header": True, }, { "name": "localized_index_json_no_cache", "path": "/templates/index.zh.json", "status": 200, "expected_cache": "no-cache", "should_have_header": True, }, # Non-matching files { "name": "html_no_header", "path": "/index.html", "status": 200, "expected_cache": None, "should_have_header": False, }, { "name": "txt_no_header", "path": "/data.txt", "status": 200, "expected_cache": None, "should_have_header": False, }, { "name": "api_endpoint_no_header", "path": "/api/endpoint", "status": 200, "expected_cache": None, "should_have_header": False, }, { "name": "pdf_no_header", "path": "/file.pdf", "status": 200, "expected_cache": None, "should_have_header": False, }, ] # Status code scenarios for images IMAGE_STATUS_SCENARIOS = [ # Success statuses get long cache {"status": 200, "expected": f"public, max-age={ONE_DAY}"}, {"status": 201, "expected": f"public, max-age={ONE_DAY}"}, {"status": 202, "expected": f"public, max-age={ONE_DAY}"}, {"status": 204, "expected": f"public, max-age={ONE_DAY}"}, {"status": 206, "expected": f"public, max-age={ONE_DAY}"}, # Permanent redirects get long cache {"status": 301, "expected": f"public, max-age={ONE_DAY}"}, {"status": 308, "expected": f"public, max-age={ONE_DAY}"}, # Temporary redirects get no cache {"status": 302, "expected": "no-cache"}, {"status": 303, "expected": "no-cache"}, {"status": 307, "expected": "no-cache"}, # 404 gets short cache {"status": 404, "expected": f"public, max-age={ONE_HOUR}"}, ] # Case sensitivity test paths CASE_SENSITIVITY_PATHS = ["/image.JPG", "/photo.PNG", "/pic.JpEg"] # Edge case test paths EDGE_CASE_PATHS = [ { "name": "query_strings_ignored", "path": "/image.jpg?v=123&size=large", "expected": f"public, max-age={ONE_DAY}", }, { "name": "multiple_dots_in_path", "path": "/image.min.jpg", "expected": f"public, max-age={ONE_DAY}", }, { "name": "nested_paths_with_images", "path": "/static/images/photo.jpg", "expected": f"public, max-age={ONE_DAY}", }, ] class TestCacheControl: """Test cache control middleware functionality""" @pytest.fixture def status_handler_factory(self): """Create a factory for handlers that return specific status codes""" def factory(status: int, headers: Dict[str, str] = None): async def handler(request): return web.Response(status=status, headers=headers or {}) return handler return factory @pytest.fixture def mock_handler(self, status_handler_factory): """Create a mock handler that returns a response with 200 status""" return status_handler_factory(200) @pytest.fixture def handler_with_existing_cache(self, status_handler_factory): """Create a handler that returns response with existing Cache-Control header""" return status_handler_factory(200, {"Cache-Control": "max-age=3600"}) async def assert_cache_header( self, response: web.Response, expected_cache: str = None, should_have_header: bool = True, ): """Helper to assert cache control headers""" if should_have_header: assert "Cache-Control" in response.headers if expected_cache: assert response.headers["Cache-Control"] == expected_cache else: assert "Cache-Control" not in response.headers # Parameterized tests @pytest.mark.parametrize("scenario", CACHE_SCENARIOS, ids=lambda x: x["name"]) async def test_cache_control_scenarios( self, scenario: Dict[str, Any], status_handler_factory ): """Test various cache control scenarios""" handler = status_handler_factory(scenario["status"]) request = make_mocked_request("GET", scenario["path"]) response = await cache_control(request, handler) assert response.status == scenario["status"] await self.assert_cache_header( response, scenario["expected_cache"], scenario["should_have_header"] ) @pytest.mark.parametrize("ext", IMG_EXTENSIONS) async def test_all_image_extensions(self, ext: str, mock_handler): """Test all defined image extensions are handled correctly""" request = make_mocked_request("GET", f"/image{ext}") response = await cache_control(request, mock_handler) assert response.status == 200 assert "Cache-Control" in response.headers assert response.headers["Cache-Control"] == f"public, max-age={ONE_DAY}" @pytest.mark.parametrize( "status_scenario", IMAGE_STATUS_SCENARIOS, ids=lambda x: f"status_{x['status']}" ) async def test_image_status_codes( self, status_scenario: Dict[str, Any], status_handler_factory ): """Test different status codes for image requests""" handler = status_handler_factory(status_scenario["status"]) request = make_mocked_request("GET", "/image.jpg") response = await cache_control(request, handler) assert response.status == status_scenario["status"] assert "Cache-Control" in response.headers assert response.headers["Cache-Control"] == status_scenario["expected"] @pytest.mark.parametrize("path", CASE_SENSITIVITY_PATHS) async def test_case_insensitive_image_extension(self, path: str, mock_handler): """Test that image extensions are matched case-insensitively""" request = make_mocked_request("GET", path) response = await cache_control(request, mock_handler) assert "Cache-Control" in response.headers assert response.headers["Cache-Control"] == f"public, max-age={ONE_DAY}" @pytest.mark.parametrize("edge_case", EDGE_CASE_PATHS, ids=lambda x: x["name"]) async def test_edge_cases(self, edge_case: Dict[str, str], mock_handler): """Test edge cases like query strings, nested paths, etc.""" request = make_mocked_request("GET", edge_case["path"]) response = await cache_control(request, mock_handler) assert "Cache-Control" in response.headers assert response.headers["Cache-Control"] == edge_case["expected"] # Header preservation tests (special cases not covered by parameterization) async def test_js_preserves_existing_headers(self, handler_with_existing_cache): """Test that .js files preserve existing Cache-Control headers""" request = make_mocked_request("GET", "/script.js") response = await cache_control(request, handler_with_existing_cache) # setdefault should preserve existing header assert response.headers["Cache-Control"] == "max-age=3600" async def test_css_preserves_existing_headers(self, handler_with_existing_cache): """Test that .css files preserve existing Cache-Control headers""" request = make_mocked_request("GET", "/styles.css") response = await cache_control(request, handler_with_existing_cache) # setdefault should preserve existing header assert response.headers["Cache-Control"] == "max-age=3600" async def test_image_preserves_existing_headers(self, status_handler_factory): """Test that image cache headers preserve existing Cache-Control""" handler = status_handler_factory(200, {"Cache-Control": "private, no-cache"}) request = make_mocked_request("GET", "/image.jpg") response = await cache_control(request, handler) # setdefault should preserve existing header assert response.headers["Cache-Control"] == "private, no-cache" async def test_304_not_modified_inherits_cache(self, status_handler_factory): """Test that 304 Not Modified doesn't set cache headers for images""" handler = status_handler_factory(304, {"Cache-Control": "max-age=7200"}) request = make_mocked_request("GET", "/not-modified.jpg") response = await cache_control(request, handler) assert response.status == 304 # Should preserve existing cache header, not override assert response.headers["Cache-Control"] == "max-age=7200"
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "tests-unit/server_test/test_cache_control.py", "license": "GNU General Public License v3.0", "lines": 228, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
Comfy-Org/ComfyUI:comfy/ldm/hunyuan3dv2_1/hunyuandit.py
import math import torch import torch.nn as nn import torch.nn.functional as F from comfy.ldm.modules.attention import optimized_attention import comfy.model_management class GELU(nn.Module): def __init__(self, dim_in: int, dim_out: int, operations, device, dtype): super().__init__() self.proj = operations.Linear(dim_in, dim_out, device = device, dtype = dtype) def gelu(self, gate: torch.Tensor) -> torch.Tensor: if gate.device.type == "mps": return F.gelu(gate.to(dtype = torch.float32)).to(dtype = gate.dtype) return F.gelu(gate) def forward(self, hidden_states): hidden_states = self.proj(hidden_states) hidden_states = self.gelu(hidden_states) return hidden_states class FeedForward(nn.Module): def __init__(self, dim: int, dim_out = None, mult: int = 4, dropout: float = 0.0, inner_dim = None, operations = None, device = None, dtype = None): super().__init__() if inner_dim is None: inner_dim = int(dim * mult) dim_out = dim_out if dim_out is not None else dim act_fn = GELU(dim, inner_dim, operations = operations, device = device, dtype = dtype) self.net = nn.ModuleList([]) self.net.append(act_fn) self.net.append(nn.Dropout(dropout)) self.net.append(operations.Linear(inner_dim, dim_out, device = device, dtype = dtype)) def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: for module in self.net: hidden_states = module(hidden_states) return hidden_states class AddAuxLoss(torch.autograd.Function): @staticmethod def forward(ctx, x, loss): # do nothing in forward (no computation) ctx.requires_aux_loss = loss.requires_grad ctx.dtype = loss.dtype return x @staticmethod def backward(ctx, grad_output): # add the aux loss gradients grad_loss = None # put the aux grad the same as the main grad loss # aux grad contributes equally if ctx.requires_aux_loss: grad_loss = torch.ones(1, dtype = ctx.dtype, device = grad_output.device) return grad_output, grad_loss class MoEGate(nn.Module): def __init__(self, embed_dim, num_experts=16, num_experts_per_tok=2, aux_loss_alpha=0.01, device = None, dtype = None): super().__init__() self.top_k = num_experts_per_tok self.n_routed_experts = num_experts self.alpha = aux_loss_alpha self.gating_dim = embed_dim self.weight = nn.Parameter(torch.empty((self.n_routed_experts, self.gating_dim), device = device, dtype = dtype)) def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: # flatten hidden states hidden_states = hidden_states.view(-1, hidden_states.size(-1)) # get logits and pass it to softmax logits = F.linear(hidden_states, comfy.model_management.cast_to(self.weight, dtype=hidden_states.dtype, device=hidden_states.device), bias = None) scores = logits.softmax(dim = -1) topk_weight, topk_idx = torch.topk(scores, k = self.top_k, dim = -1, sorted = False) if self.training and self.alpha > 0.0: scores_for_aux = scores # used bincount instead of one hot encoding counts = torch.bincount(topk_idx.view(-1), minlength = self.n_routed_experts).float() ce = counts / topk_idx.numel() # normalized expert usage # mean expert score Pi = scores_for_aux.mean(0) # expert balance loss aux_loss = (Pi * ce * self.n_routed_experts).sum() * self.alpha else: aux_loss = None return topk_idx, topk_weight, aux_loss class MoEBlock(nn.Module): def __init__(self, dim, num_experts: int = 6, moe_top_k: int = 2, dropout: float = 0.0, ff_inner_dim: int = None, operations = None, device = None, dtype = None): super().__init__() self.moe_top_k = moe_top_k self.num_experts = num_experts self.experts = nn.ModuleList([ FeedForward(dim, dropout = dropout, inner_dim = ff_inner_dim, operations = operations, device = device, dtype = dtype) for _ in range(num_experts) ]) self.gate = MoEGate(dim, num_experts = num_experts, num_experts_per_tok = moe_top_k, device = device, dtype = dtype) self.shared_experts = FeedForward(dim, dropout = dropout, inner_dim = ff_inner_dim, operations = operations, device = device, dtype = dtype) def forward(self, hidden_states) -> torch.Tensor: identity = hidden_states orig_shape = hidden_states.shape topk_idx, topk_weight, aux_loss = self.gate(hidden_states) hidden_states = hidden_states.view(-1, hidden_states.shape[-1]) flat_topk_idx = topk_idx.view(-1) if self.training: hidden_states = hidden_states.repeat_interleave(self.moe_top_k, dim = 0) y = torch.empty_like(hidden_states, dtype = hidden_states.dtype) for i, expert in enumerate(self.experts): tmp = expert(hidden_states[flat_topk_idx == i]) y[flat_topk_idx == i] = tmp.to(hidden_states.dtype) y = (y.view(*topk_weight.shape, -1) * topk_weight.unsqueeze(-1)).sum(dim = 1) y = y.view(*orig_shape) y = AddAuxLoss.apply(y, aux_loss) else: y = self.moe_infer(hidden_states, flat_expert_indices = flat_topk_idx,flat_expert_weights = topk_weight.view(-1, 1)).view(*orig_shape) y = y + self.shared_experts(identity) return y @torch.no_grad() def moe_infer(self, x, flat_expert_indices, flat_expert_weights): expert_cache = torch.zeros_like(x) idxs = flat_expert_indices.argsort() # no need for .numpy().cpu() here tokens_per_expert = flat_expert_indices.bincount().cumsum(0) token_idxs = idxs // self.moe_top_k for i, end_idx in enumerate(tokens_per_expert): start_idx = 0 if i == 0 else tokens_per_expert[i-1] if start_idx == end_idx: continue expert = self.experts[i] exp_token_idx = token_idxs[start_idx:end_idx] expert_tokens = x[exp_token_idx] expert_out = expert(expert_tokens) expert_out.mul_(flat_expert_weights[idxs[start_idx:end_idx]]) # use index_add_ with a 1-D index tensor directly avoids building a large [N, D] index map and extra memcopy required by scatter_reduce_ # + avoid dtype conversion expert_cache.index_add_(0, exp_token_idx, expert_out) return expert_cache class Timesteps(nn.Module): def __init__(self, num_channels: int, downscale_freq_shift: float = 0.0, scale: float = 1.0, max_period: int = 10000): super().__init__() self.num_channels = num_channels half_dim = num_channels // 2 # precompute the “inv_freq” vector once exponent = -math.log(max_period) * torch.arange( half_dim, dtype=torch.float32 ) / (half_dim - downscale_freq_shift) inv_freq = torch.exp(exponent) # pad if num_channels % 2 == 1: # we’ll pad a zero at the end of the cos-half inv_freq = torch.cat([inv_freq, inv_freq.new_zeros(1)]) # register to buffer so it moves with the device self.register_buffer("inv_freq", inv_freq, persistent = False) self.scale = scale def forward(self, timesteps: torch.Tensor): x = timesteps.float().unsqueeze(1) * self.inv_freq.to(timesteps.device).unsqueeze(0) # fused CUDA kernels for sin and cos sin_emb = x.sin() cos_emb = x.cos() emb = torch.cat([sin_emb, cos_emb], dim = 1) # scale factor if self.scale != 1.0: emb = emb * self.scale # If we padded inv_freq for odd, emb is already wide enough; otherwise: if emb.shape[1] > self.num_channels: emb = emb[:, :self.num_channels] return emb class TimestepEmbedder(nn.Module): def __init__(self, hidden_size, frequency_embedding_size = 256, cond_proj_dim = None, operations = None, device = None, dtype = None): super().__init__() self.mlp = nn.Sequential( operations.Linear(hidden_size, frequency_embedding_size, bias=True, device = device, dtype = dtype), nn.GELU(), operations.Linear(frequency_embedding_size, hidden_size, bias=True, device = device, dtype = dtype), ) self.frequency_embedding_size = frequency_embedding_size if cond_proj_dim is not None: self.cond_proj = operations.Linear(cond_proj_dim, frequency_embedding_size, bias=False, device = device, dtype = dtype) self.time_embed = Timesteps(hidden_size) def forward(self, timesteps, condition): timestep_embed = self.time_embed(timesteps).type(self.mlp[0].weight.dtype) if condition is not None: cond_embed = self.cond_proj(condition) timestep_embed = timestep_embed + cond_embed time_conditioned = self.mlp(timestep_embed) # for broadcasting with image tokens return time_conditioned.unsqueeze(1) class MLP(nn.Module): def __init__(self, *, width: int, operations = None, device = None, dtype = None): super().__init__() self.width = width self.fc1 = operations.Linear(width, width * 4, device = device, dtype = dtype) self.fc2 = operations.Linear(width * 4, width, device = device, dtype = dtype) self.gelu = nn.GELU() def forward(self, x): return self.fc2(self.gelu(self.fc1(x))) class CrossAttention(nn.Module): def __init__( self, qdim, kdim, num_heads, qkv_bias=True, qk_norm=False, norm_layer=nn.LayerNorm, use_fp16: bool = False, operations = None, dtype = None, device = None, **kwargs, ): super().__init__() self.qdim = qdim self.kdim = kdim self.num_heads = num_heads self.head_dim = self.qdim // num_heads self.scale = self.head_dim ** -0.5 self.to_q = operations.Linear(qdim, qdim, bias=qkv_bias, device = device, dtype = dtype) self.to_k = operations.Linear(kdim, qdim, bias=qkv_bias, device = device, dtype = dtype) self.to_v = operations.Linear(kdim, qdim, bias=qkv_bias, device = device, dtype = dtype) if use_fp16: eps = 1.0 / 65504 else: eps = 1e-6 if norm_layer == nn.LayerNorm: norm_layer = operations.LayerNorm else: norm_layer = operations.RMSNorm self.q_norm = norm_layer(self.head_dim, elementwise_affine=True, eps = eps, device = device, dtype = dtype) if qk_norm else nn.Identity() self.k_norm = norm_layer(self.head_dim, elementwise_affine=True, eps = eps, device = device, dtype = dtype) if qk_norm else nn.Identity() self.out_proj = operations.Linear(qdim, qdim, bias=True, device = device, dtype = dtype) def forward(self, x, y): b, s1, _ = x.shape _, s2, _ = y.shape y = y.to(next(self.to_k.parameters()).dtype) q = self.to_q(x) k = self.to_k(y) v = self.to_v(y) kv = torch.cat((k, v), dim=-1) split_size = kv.shape[-1] // self.num_heads // 2 kv = kv.view(1, -1, self.num_heads, split_size * 2) k, v = torch.split(kv, split_size, dim=-1) q = q.view(b, s1, self.num_heads, self.head_dim) k = k.view(b, s2, self.num_heads, self.head_dim) v = v.reshape(b, s2, self.num_heads * self.head_dim) q = self.q_norm(q) k = self.k_norm(k) x = optimized_attention( q.reshape(b, s1, self.num_heads * self.head_dim), k.reshape(b, s2, self.num_heads * self.head_dim), v, heads=self.num_heads, ) out = self.out_proj(x) return out class Attention(nn.Module): def __init__( self, dim, num_heads, qkv_bias = True, qk_norm = False, norm_layer = nn.LayerNorm, use_fp16: bool = False, operations = None, device = None, dtype = None ): super().__init__() self.dim = dim self.num_heads = num_heads self.head_dim = self.dim // num_heads self.scale = self.head_dim ** -0.5 self.to_q = operations.Linear(dim, dim, bias = qkv_bias, device = device, dtype = dtype) self.to_k = operations.Linear(dim, dim, bias = qkv_bias, device = device, dtype = dtype) self.to_v = operations.Linear(dim, dim, bias = qkv_bias, device = device, dtype = dtype) if use_fp16: eps = 1.0 / 65504 else: eps = 1e-6 if norm_layer == nn.LayerNorm: norm_layer = operations.LayerNorm else: norm_layer = operations.RMSNorm self.q_norm = norm_layer(self.head_dim, elementwise_affine=True, eps = eps, device = device, dtype = dtype) if qk_norm else nn.Identity() self.k_norm = norm_layer(self.head_dim, elementwise_affine=True, eps = eps, device = device, dtype = dtype) if qk_norm else nn.Identity() self.out_proj = operations.Linear(dim, dim, device = device, dtype = dtype) def forward(self, x): B, N, _ = x.shape query = self.to_q(x) key = self.to_k(x) value = self.to_v(x) qkv_combined = torch.cat((query, key, value), dim=-1) split_size = qkv_combined.shape[-1] // self.num_heads // 3 qkv = qkv_combined.view(1, -1, self.num_heads, split_size * 3) query, key, value = torch.split(qkv, split_size, dim=-1) query = query.reshape(B, N, self.num_heads, self.head_dim) key = key.reshape(B, N, self.num_heads, self.head_dim) value = value.reshape(B, N, self.num_heads * self.head_dim) query = self.q_norm(query) key = self.k_norm(key) x = optimized_attention( query.reshape(B, N, self.num_heads * self.head_dim), key.reshape(B, N, self.num_heads * self.head_dim), value, heads=self.num_heads, ) x = self.out_proj(x) return x class HunYuanDiTBlock(nn.Module): def __init__( self, hidden_size, c_emb_size, num_heads, text_states_dim=1024, qk_norm=False, norm_layer=nn.LayerNorm, qk_norm_layer=True, qkv_bias=True, skip_connection=True, timested_modulate=False, use_moe: bool = False, num_experts: int = 8, moe_top_k: int = 2, use_fp16: bool = False, operations = None, device = None, dtype = None ): super().__init__() # eps can't be 1e-6 in fp16 mode because of numerical stability issues if use_fp16: eps = 1.0 / 65504 else: eps = 1e-6 self.norm1 = norm_layer(hidden_size, elementwise_affine = True, eps = eps, device = device, dtype = dtype) self.attn1 = Attention(hidden_size, num_heads=num_heads, qkv_bias=qkv_bias, qk_norm=qk_norm, norm_layer=qk_norm_layer, use_fp16 = use_fp16, device = device, dtype = dtype, operations = operations) self.norm2 = norm_layer(hidden_size, elementwise_affine = True, eps = eps, device = device, dtype = dtype) self.timested_modulate = timested_modulate if self.timested_modulate: self.default_modulation = nn.Sequential( nn.SiLU(), operations.Linear(c_emb_size, hidden_size, bias=True, device = device, dtype = dtype) ) self.attn2 = CrossAttention(hidden_size, text_states_dim, num_heads=num_heads, qkv_bias=qkv_bias, qk_norm=qk_norm, norm_layer=qk_norm_layer, use_fp16 = use_fp16, device = device, dtype = dtype, operations = operations) self.norm3 = norm_layer(hidden_size, elementwise_affine = True, eps = eps, device = device, dtype = dtype) if skip_connection: self.skip_norm = norm_layer(hidden_size, elementwise_affine = True, eps = eps, device = device, dtype = dtype) self.skip_linear = operations.Linear(2 * hidden_size, hidden_size, device = device, dtype = dtype) else: self.skip_linear = None self.use_moe = use_moe if self.use_moe: self.moe = MoEBlock( hidden_size, num_experts = num_experts, moe_top_k = moe_top_k, dropout = 0.0, ff_inner_dim = int(hidden_size * 4.0), device = device, dtype = dtype, operations = operations ) else: self.mlp = MLP(width=hidden_size, operations=operations, device = device, dtype = dtype) def forward(self, hidden_states, conditioning=None, text_states=None, skip_tensor=None): if self.skip_linear is not None: combined = torch.cat([skip_tensor, hidden_states], dim=-1) hidden_states = self.skip_linear(combined) hidden_states = self.skip_norm(hidden_states) # self attention if self.timested_modulate: modulation_shift = self.default_modulation(conditioning).unsqueeze(dim=1) hidden_states = hidden_states + modulation_shift self_attn_out = self.attn1(self.norm1(hidden_states)) hidden_states = hidden_states + self_attn_out # cross attention hidden_states = hidden_states + self.attn2(self.norm2(hidden_states), text_states) # MLP Layer mlp_input = self.norm3(hidden_states) if self.use_moe: hidden_states = hidden_states + self.moe(mlp_input) else: hidden_states = hidden_states + self.mlp(mlp_input) return hidden_states class FinalLayer(nn.Module): def __init__(self, final_hidden_size, out_channels, operations, use_fp16: bool = False, device = None, dtype = None): super().__init__() if use_fp16: eps = 1.0 / 65504 else: eps = 1e-6 self.norm_final = operations.LayerNorm(final_hidden_size, elementwise_affine = True, eps = eps, device = device, dtype = dtype) self.linear = operations.Linear(final_hidden_size, out_channels, bias = True, device = device, dtype = dtype) def forward(self, x): x = self.norm_final(x) x = x[:, 1:] x = self.linear(x) return x class HunYuanDiTPlain(nn.Module): # init with the defaults values from https://huggingface.co/tencent/Hunyuan3D-2.1/blob/main/hunyuan3d-dit-v2-1/config.yaml def __init__( self, in_channels: int = 64, hidden_size: int = 2048, context_dim: int = 1024, depth: int = 21, num_heads: int = 16, qk_norm: bool = True, qkv_bias: bool = False, num_moe_layers: int = 6, guidance_cond_proj_dim = 2048, norm_type = 'layer', num_experts: int = 8, moe_top_k: int = 2, use_fp16: bool = False, dtype = None, device = None, operations = None, **kwargs ): self.dtype = dtype super().__init__() self.depth = depth self.in_channels = in_channels self.out_channels = in_channels self.num_heads = num_heads self.hidden_size = hidden_size norm = operations.LayerNorm if norm_type == 'layer' else operations.RMSNorm qk_norm = operations.RMSNorm self.context_dim = context_dim self.guidance_cond_proj_dim = guidance_cond_proj_dim self.x_embedder = operations.Linear(in_channels, hidden_size, bias = True, device = device, dtype = dtype) self.t_embedder = TimestepEmbedder(hidden_size, hidden_size * 4, cond_proj_dim = guidance_cond_proj_dim, device = device, dtype = dtype, operations = operations) # HUnYuanDiT Blocks self.blocks = nn.ModuleList([ HunYuanDiTBlock(hidden_size=hidden_size, c_emb_size=hidden_size, num_heads=num_heads, text_states_dim=context_dim, qk_norm=qk_norm, norm_layer = norm, qk_norm_layer = qk_norm, skip_connection=layer > depth // 2, qkv_bias=qkv_bias, use_moe=True if depth - layer <= num_moe_layers else False, num_experts=num_experts, moe_top_k=moe_top_k, use_fp16 = use_fp16, device = device, dtype = dtype, operations = operations) for layer in range(depth) ]) self.depth = depth self.final_layer = FinalLayer(hidden_size, self.out_channels, use_fp16 = use_fp16, operations = operations, device = device, dtype = dtype) def forward(self, x, t, context, transformer_options = {}, **kwargs): x = x.movedim(-1, -2) uncond_emb, cond_emb = context.chunk(2, dim = 0) context = torch.cat([cond_emb, uncond_emb], dim = 0) main_condition = context t = 1.0 - t time_embedded = self.t_embedder(t, condition = kwargs.get('guidance_cond')) x = x.to(dtype = next(self.x_embedder.parameters()).dtype) x_embedded = self.x_embedder(x) combined = torch.cat([time_embedded, x_embedded], dim=1) def block_wrap(args): return block( args["x"], args["t"], args["cond"], skip_tensor=args.get("skip"),) skip_stack = [] patches_replace = transformer_options.get("patches_replace", {}) blocks_replace = patches_replace.get("dit", {}) for idx, block in enumerate(self.blocks): if idx <= self.depth // 2: skip_input = None else: skip_input = skip_stack.pop() if ("block", idx) in blocks_replace: combined = blocks_replace[("block", idx)]( { "x": combined, "t": time_embedded, "cond": main_condition, "skip": skip_input, }, {"original_block": block_wrap}, ) else: combined = block(combined, time_embedded, main_condition, skip_tensor=skip_input) if idx < self.depth // 2: skip_stack.append(combined) output = self.final_layer(combined) output = output.movedim(-2, -1) * (-1.0) cond_emb, uncond_emb = output.chunk(2, dim = 0) return torch.cat([uncond_emb, cond_emb])
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/ldm/hunyuan3dv2_1/hunyuandit.py", "license": "GNU General Public License v3.0", "lines": 490, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy/audio_encoders/audio_encoders.py
from .wav2vec2 import Wav2Vec2Model from .whisper import WhisperLargeV3 import comfy.model_management import comfy.ops import comfy.utils import logging import torchaudio class AudioEncoderModel(): def __init__(self, config): self.load_device = comfy.model_management.text_encoder_device() offload_device = comfy.model_management.text_encoder_offload_device() self.dtype = comfy.model_management.text_encoder_dtype(self.load_device) model_type = config.pop("model_type") model_config = dict(config) model_config.update({ "dtype": self.dtype, "device": offload_device, "operations": comfy.ops.manual_cast }) if model_type == "wav2vec2": self.model = Wav2Vec2Model(**model_config) elif model_type == "whisper3": self.model = WhisperLargeV3(**model_config) self.model.eval() self.patcher = comfy.model_patcher.CoreModelPatcher(self.model, load_device=self.load_device, offload_device=offload_device) self.model_sample_rate = 16000 def load_sd(self, sd): return self.model.load_state_dict(sd, strict=False, assign=self.patcher.is_dynamic()) def get_sd(self): return self.model.state_dict() def encode_audio(self, audio, sample_rate): comfy.model_management.load_model_gpu(self.patcher) audio = torchaudio.functional.resample(audio, sample_rate, self.model_sample_rate) out, all_layers = self.model(audio.to(self.load_device)) outputs = {} outputs["encoded_audio"] = out outputs["encoded_audio_all_layers"] = all_layers outputs["audio_samples"] = audio.shape[2] return outputs def load_audio_encoder_from_sd(sd, prefix=""): sd = comfy.utils.state_dict_prefix_replace(sd, {"wav2vec2.": ""}) if "encoder.layer_norm.bias" in sd: #wav2vec2 embed_dim = sd["encoder.layer_norm.bias"].shape[0] if embed_dim == 1024:# large config = { "model_type": "wav2vec2", "embed_dim": 1024, "num_heads": 16, "num_layers": 24, "conv_norm": True, "conv_bias": True, "do_normalize": True, "do_stable_layer_norm": True } elif embed_dim == 768: # base config = { "model_type": "wav2vec2", "embed_dim": 768, "num_heads": 12, "num_layers": 12, "conv_norm": False, "conv_bias": False, "do_normalize": False, # chinese-wav2vec2-base has this False "do_stable_layer_norm": False } else: raise RuntimeError("ERROR: audio encoder file is invalid or unsupported embed_dim: {}".format(embed_dim)) elif "model.encoder.embed_positions.weight" in sd: sd = comfy.utils.state_dict_prefix_replace(sd, {"model.": ""}) config = { "model_type": "whisper3", } else: raise RuntimeError("ERROR: audio encoder not supported.") audio_encoder = AudioEncoderModel(config) m, u = audio_encoder.load_sd(sd) if len(m) > 0: logging.warning("missing audio encoder: {}".format(m)) if len(u) > 0: logging.warning("unexpected audio encoder: {}".format(u)) return audio_encoder
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/audio_encoders/audio_encoders.py", "license": "GNU General Public License v3.0", "lines": 81, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy/audio_encoders/wav2vec2.py
import torch import torch.nn as nn from comfy.ldm.modules.attention import optimized_attention_masked class LayerNormConv(nn.Module): def __init__(self, in_channels, out_channels, kernel_size, stride, bias=False, dtype=None, device=None, operations=None): super().__init__() self.conv = operations.Conv1d(in_channels, out_channels, kernel_size=kernel_size, stride=stride, bias=bias, device=device, dtype=dtype) self.layer_norm = operations.LayerNorm(out_channels, elementwise_affine=True, device=device, dtype=dtype) def forward(self, x): x = self.conv(x) return torch.nn.functional.gelu(self.layer_norm(x.transpose(-2, -1)).transpose(-2, -1)) class LayerGroupNormConv(nn.Module): def __init__(self, in_channels, out_channels, kernel_size, stride, bias=False, dtype=None, device=None, operations=None): super().__init__() self.conv = operations.Conv1d(in_channels, out_channels, kernel_size=kernel_size, stride=stride, bias=bias, device=device, dtype=dtype) self.layer_norm = operations.GroupNorm(num_groups=out_channels, num_channels=out_channels, affine=True, device=device, dtype=dtype) def forward(self, x): x = self.conv(x) return torch.nn.functional.gelu(self.layer_norm(x)) class ConvNoNorm(nn.Module): def __init__(self, in_channels, out_channels, kernel_size, stride, bias=False, dtype=None, device=None, operations=None): super().__init__() self.conv = operations.Conv1d(in_channels, out_channels, kernel_size=kernel_size, stride=stride, bias=bias, device=device, dtype=dtype) def forward(self, x): x = self.conv(x) return torch.nn.functional.gelu(x) class ConvFeatureEncoder(nn.Module): def __init__(self, conv_dim, conv_bias=False, conv_norm=True, dtype=None, device=None, operations=None): super().__init__() if conv_norm: self.conv_layers = nn.ModuleList([ LayerNormConv(1, conv_dim, kernel_size=10, stride=5, bias=True, device=device, dtype=dtype, operations=operations), LayerNormConv(conv_dim, conv_dim, kernel_size=3, stride=2, bias=conv_bias, device=device, dtype=dtype, operations=operations), LayerNormConv(conv_dim, conv_dim, kernel_size=3, stride=2, bias=conv_bias, device=device, dtype=dtype, operations=operations), LayerNormConv(conv_dim, conv_dim, kernel_size=3, stride=2, bias=conv_bias, device=device, dtype=dtype, operations=operations), LayerNormConv(conv_dim, conv_dim, kernel_size=3, stride=2, bias=conv_bias, device=device, dtype=dtype, operations=operations), LayerNormConv(conv_dim, conv_dim, kernel_size=2, stride=2, bias=conv_bias, device=device, dtype=dtype, operations=operations), LayerNormConv(conv_dim, conv_dim, kernel_size=2, stride=2, bias=conv_bias, device=device, dtype=dtype, operations=operations), ]) else: self.conv_layers = nn.ModuleList([ LayerGroupNormConv(1, conv_dim, kernel_size=10, stride=5, bias=conv_bias, device=device, dtype=dtype, operations=operations), ConvNoNorm(conv_dim, conv_dim, kernel_size=3, stride=2, bias=conv_bias, device=device, dtype=dtype, operations=operations), ConvNoNorm(conv_dim, conv_dim, kernel_size=3, stride=2, bias=conv_bias, device=device, dtype=dtype, operations=operations), ConvNoNorm(conv_dim, conv_dim, kernel_size=3, stride=2, bias=conv_bias, device=device, dtype=dtype, operations=operations), ConvNoNorm(conv_dim, conv_dim, kernel_size=3, stride=2, bias=conv_bias, device=device, dtype=dtype, operations=operations), ConvNoNorm(conv_dim, conv_dim, kernel_size=2, stride=2, bias=conv_bias, device=device, dtype=dtype, operations=operations), ConvNoNorm(conv_dim, conv_dim, kernel_size=2, stride=2, bias=conv_bias, device=device, dtype=dtype, operations=operations), ]) def forward(self, x): x = x.unsqueeze(1) for conv in self.conv_layers: x = conv(x) return x.transpose(1, 2) class FeatureProjection(nn.Module): def __init__(self, conv_dim, embed_dim, dtype=None, device=None, operations=None): super().__init__() self.layer_norm = operations.LayerNorm(conv_dim, eps=1e-05, device=device, dtype=dtype) self.projection = operations.Linear(conv_dim, embed_dim, device=device, dtype=dtype) def forward(self, x): x = self.layer_norm(x) x = self.projection(x) return x class PositionalConvEmbedding(nn.Module): def __init__(self, embed_dim=768, kernel_size=128, groups=16): super().__init__() self.conv = nn.Conv1d( embed_dim, embed_dim, kernel_size=kernel_size, padding=kernel_size // 2, groups=groups, ) self.conv = torch.nn.utils.parametrizations.weight_norm(self.conv, name="weight", dim=2) self.activation = nn.GELU() def forward(self, x): x = x.transpose(1, 2) x = self.conv(x)[:, :, :-1] x = self.activation(x) x = x.transpose(1, 2) return x class TransformerEncoder(nn.Module): def __init__( self, embed_dim=768, num_heads=12, num_layers=12, mlp_ratio=4.0, do_stable_layer_norm=True, dtype=None, device=None, operations=None ): super().__init__() self.pos_conv_embed = PositionalConvEmbedding(embed_dim=embed_dim) self.layers = nn.ModuleList([ TransformerEncoderLayer( embed_dim=embed_dim, num_heads=num_heads, mlp_ratio=mlp_ratio, do_stable_layer_norm=do_stable_layer_norm, device=device, dtype=dtype, operations=operations ) for _ in range(num_layers) ]) self.layer_norm = operations.LayerNorm(embed_dim, eps=1e-05, device=device, dtype=dtype) self.do_stable_layer_norm = do_stable_layer_norm def forward(self, x, mask=None): x = x + self.pos_conv_embed(x) all_x = () if not self.do_stable_layer_norm: x = self.layer_norm(x) for layer in self.layers: all_x += (x,) x = layer(x, mask) if self.do_stable_layer_norm: x = self.layer_norm(x) all_x += (x,) return x, all_x class Attention(nn.Module): def __init__(self, embed_dim, num_heads, bias=True, dtype=None, device=None, operations=None): super().__init__() self.embed_dim = embed_dim self.num_heads = num_heads self.head_dim = embed_dim // num_heads self.k_proj = operations.Linear(embed_dim, embed_dim, bias=bias, device=device, dtype=dtype) self.v_proj = operations.Linear(embed_dim, embed_dim, bias=bias, device=device, dtype=dtype) self.q_proj = operations.Linear(embed_dim, embed_dim, bias=bias, device=device, dtype=dtype) self.out_proj = operations.Linear(embed_dim, embed_dim, bias=bias, device=device, dtype=dtype) def forward(self, x, mask=None): assert (mask is None) # TODO? q = self.q_proj(x) k = self.k_proj(x) v = self.v_proj(x) out = optimized_attention_masked(q, k, v, self.num_heads) return self.out_proj(out) class FeedForward(nn.Module): def __init__(self, embed_dim, mlp_ratio, dtype=None, device=None, operations=None): super().__init__() self.intermediate_dense = operations.Linear(embed_dim, int(embed_dim * mlp_ratio), device=device, dtype=dtype) self.output_dense = operations.Linear(int(embed_dim * mlp_ratio), embed_dim, device=device, dtype=dtype) def forward(self, x): x = self.intermediate_dense(x) x = torch.nn.functional.gelu(x) x = self.output_dense(x) return x class TransformerEncoderLayer(nn.Module): def __init__( self, embed_dim=768, num_heads=12, mlp_ratio=4.0, do_stable_layer_norm=True, dtype=None, device=None, operations=None ): super().__init__() self.attention = Attention(embed_dim, num_heads, device=device, dtype=dtype, operations=operations) self.layer_norm = operations.LayerNorm(embed_dim, device=device, dtype=dtype) self.feed_forward = FeedForward(embed_dim, mlp_ratio, device=device, dtype=dtype, operations=operations) self.final_layer_norm = operations.LayerNorm(embed_dim, device=device, dtype=dtype) self.do_stable_layer_norm = do_stable_layer_norm def forward(self, x, mask=None): residual = x if self.do_stable_layer_norm: x = self.layer_norm(x) x = self.attention(x, mask=mask) x = residual + x if not self.do_stable_layer_norm: x = self.layer_norm(x) return self.final_layer_norm(x + self.feed_forward(x)) else: return x + self.feed_forward(self.final_layer_norm(x)) class Wav2Vec2Model(nn.Module): """Complete Wav2Vec 2.0 model.""" def __init__( self, embed_dim=1024, final_dim=256, num_heads=16, num_layers=24, conv_norm=True, conv_bias=True, do_normalize=True, do_stable_layer_norm=True, dtype=None, device=None, operations=None ): super().__init__() conv_dim = 512 self.feature_extractor = ConvFeatureEncoder(conv_dim, conv_norm=conv_norm, conv_bias=conv_bias, device=device, dtype=dtype, operations=operations) self.feature_projection = FeatureProjection(conv_dim, embed_dim, device=device, dtype=dtype, operations=operations) self.masked_spec_embed = nn.Parameter(torch.empty(embed_dim, device=device, dtype=dtype)) self.do_normalize = do_normalize self.encoder = TransformerEncoder( embed_dim=embed_dim, num_heads=num_heads, num_layers=num_layers, do_stable_layer_norm=do_stable_layer_norm, device=device, dtype=dtype, operations=operations ) def forward(self, x, mask_time_indices=None, return_dict=False): x = torch.mean(x, dim=1) if self.do_normalize: x = (x - x.mean()) / torch.sqrt(x.var() + 1e-7) features = self.feature_extractor(x) features = self.feature_projection(features) batch_size, seq_len, _ = features.shape x, all_x = self.encoder(features) return x, all_x
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/audio_encoders/wav2vec2.py", "license": "GNU General Public License v3.0", "lines": 206, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy_extras/nodes_audio_encoder.py
import folder_paths import comfy.audio_encoders.audio_encoders import comfy.utils from typing_extensions import override from comfy_api.latest import ComfyExtension, io class AudioEncoderLoader(io.ComfyNode): @classmethod def define_schema(cls) -> io.Schema: return io.Schema( node_id="AudioEncoderLoader", category="loaders", inputs=[ io.Combo.Input( "audio_encoder_name", options=folder_paths.get_filename_list("audio_encoders"), ), ], outputs=[io.AudioEncoder.Output()], ) @classmethod def execute(cls, audio_encoder_name) -> io.NodeOutput: audio_encoder_name = folder_paths.get_full_path_or_raise("audio_encoders", audio_encoder_name) sd = comfy.utils.load_torch_file(audio_encoder_name, safe_load=True) audio_encoder = comfy.audio_encoders.audio_encoders.load_audio_encoder_from_sd(sd) if audio_encoder is None: raise RuntimeError("ERROR: audio encoder file is invalid and does not contain a valid model.") return io.NodeOutput(audio_encoder) class AudioEncoderEncode(io.ComfyNode): @classmethod def define_schema(cls) -> io.Schema: return io.Schema( node_id="AudioEncoderEncode", category="conditioning", inputs=[ io.AudioEncoder.Input("audio_encoder"), io.Audio.Input("audio"), ], outputs=[io.AudioEncoderOutput.Output()], ) @classmethod def execute(cls, audio_encoder, audio) -> io.NodeOutput: output = audio_encoder.encode_audio(audio["waveform"], audio["sample_rate"]) return io.NodeOutput(output) class AudioEncoder(ComfyExtension): @override async def get_node_list(self) -> list[type[io.ComfyNode]]: return [ AudioEncoderLoader, AudioEncoderEncode, ] async def comfy_entrypoint() -> AudioEncoder: return AudioEncoder()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_extras/nodes_audio_encoder.py", "license": "GNU General Public License v3.0", "lines": 52, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy_extras/nodes_easycache.py
from __future__ import annotations from typing import TYPE_CHECKING, Union from comfy_api.latest import io, ComfyExtension import comfy.patcher_extension import logging import torch import comfy.model_patcher if TYPE_CHECKING: from uuid import UUID def _extract_tensor(data, output_channels): """Extract tensor from data, handling both single tensors and lists.""" if isinstance(data, list): # LTX2 AV tensors: [video, audio] return data[0][:, :output_channels], data[1][:, :output_channels] return data[:, :output_channels], None def easycache_forward_wrapper(executor, *args, **kwargs): # get values from args transformer_options: dict[str] = args[-1] if not isinstance(transformer_options, dict): transformer_options = kwargs.get("transformer_options") if not transformer_options: transformer_options = args[-2] easycache: EasyCacheHolder = transformer_options["easycache"] x, ax = _extract_tensor(args[0], easycache.output_channels) sigmas = transformer_options["sigmas"] uuids = transformer_options["uuids"] if sigmas is not None and easycache.is_past_end_timestep(sigmas): return executor(*args, **kwargs) # prepare next x_prev has_first_cond_uuid = easycache.has_first_cond_uuid(uuids) next_x_prev = x input_change = None do_easycache = easycache.should_do_easycache(sigmas) if do_easycache: easycache.check_metadata(x) # if there isn't a cache diff for current conds, we cannot skip this step can_apply_cache_diff = easycache.can_apply_cache_diff(uuids) # if first cond marked this step for skipping, skip it and use appropriate cached values if easycache.skip_current_step and can_apply_cache_diff: if easycache.verbose: logging.info(f"EasyCache [verbose] - was marked to skip this step by {easycache.first_cond_uuid}. Present uuids: {uuids}") result = easycache.apply_cache_diff(x, uuids) if ax is not None: result_audio = easycache.apply_cache_diff(ax, uuids, is_audio=True) return [result, result_audio] return result if easycache.initial_step: easycache.first_cond_uuid = uuids[0] has_first_cond_uuid = easycache.has_first_cond_uuid(uuids) easycache.initial_step = False if has_first_cond_uuid: if easycache.has_x_prev_subsampled(): input_change = (easycache.subsample(x, uuids, clone=False) - easycache.x_prev_subsampled).flatten().abs().mean() if easycache.has_output_prev_norm() and easycache.has_relative_transformation_rate(): approx_output_change_rate = (easycache.relative_transformation_rate * input_change) / easycache.output_prev_norm easycache.cumulative_change_rate += approx_output_change_rate if easycache.cumulative_change_rate < easycache.reuse_threshold and can_apply_cache_diff: if easycache.verbose: logging.info(f"EasyCache [verbose] - skipping step; cumulative_change_rate: {easycache.cumulative_change_rate}, reuse_threshold: {easycache.reuse_threshold}") # other conds should also skip this step, and instead use their cached values easycache.skip_current_step = True result = easycache.apply_cache_diff(x, uuids) if ax is not None: result_audio = easycache.apply_cache_diff(ax, uuids, is_audio=True) return [result, result_audio] return result else: if easycache.verbose: logging.info(f"EasyCache [verbose] - NOT skipping step; cumulative_change_rate: {easycache.cumulative_change_rate}, reuse_threshold: {easycache.reuse_threshold}") easycache.cumulative_change_rate = 0.0 full_output: torch.Tensor = executor(*args, **kwargs) output, audio_output = _extract_tensor(full_output, easycache.output_channels) if has_first_cond_uuid and easycache.has_output_prev_norm(): output_change = (easycache.subsample(output, uuids, clone=False) - easycache.output_prev_subsampled).flatten().abs().mean() if easycache.verbose: output_change_rate = output_change / easycache.output_prev_norm easycache.output_change_rates.append(output_change_rate.item()) if easycache.has_relative_transformation_rate(): approx_output_change_rate = (easycache.relative_transformation_rate * input_change) / easycache.output_prev_norm easycache.approx_output_change_rates.append(approx_output_change_rate.item()) if easycache.verbose: logging.info(f"EasyCache [verbose] - approx_output_change_rate: {approx_output_change_rate}") if input_change is not None: easycache.relative_transformation_rate = output_change / input_change if easycache.verbose: logging.info(f"EasyCache [verbose] - output_change_rate: {output_change_rate}") # TODO: allow cache_diff to be offloaded easycache.update_cache_diff(output, next_x_prev, uuids) if audio_output is not None: easycache.update_cache_diff(audio_output, ax, uuids, is_audio=True) if has_first_cond_uuid: easycache.x_prev_subsampled = easycache.subsample(next_x_prev, uuids) easycache.output_prev_subsampled = easycache.subsample(output, uuids) easycache.output_prev_norm = output.flatten().abs().mean() if easycache.verbose: logging.info(f"EasyCache [verbose] - x_prev_subsampled: {easycache.x_prev_subsampled.shape}") return full_output def lazycache_predict_noise_wrapper(executor, *args, **kwargs): # get values from args timestep: float = args[1] model_options: dict[str] = args[2] easycache: LazyCacheHolder = model_options["transformer_options"]["easycache"] if easycache.is_past_end_timestep(timestep): return executor(*args, **kwargs) x: torch.Tensor = args[0][:, :easycache.output_channels] # prepare next x_prev next_x_prev = x input_change = None do_easycache = easycache.should_do_easycache(timestep) if do_easycache: easycache.check_metadata(x) if easycache.has_x_prev_subsampled(): if easycache.has_x_prev_subsampled(): input_change = (easycache.subsample(x, clone=False) - easycache.x_prev_subsampled).flatten().abs().mean() if easycache.has_output_prev_norm() and easycache.has_relative_transformation_rate(): approx_output_change_rate = (easycache.relative_transformation_rate * input_change) / easycache.output_prev_norm easycache.cumulative_change_rate += approx_output_change_rate if easycache.cumulative_change_rate < easycache.reuse_threshold: if easycache.verbose: logging.info(f"LazyCache [verbose] - skipping step; cumulative_change_rate: {easycache.cumulative_change_rate}, reuse_threshold: {easycache.reuse_threshold}") # other conds should also skip this step, and instead use their cached values easycache.skip_current_step = True return easycache.apply_cache_diff(x) else: if easycache.verbose: logging.info(f"LazyCache [verbose] - NOT skipping step; cumulative_change_rate: {easycache.cumulative_change_rate}, reuse_threshold: {easycache.reuse_threshold}") easycache.cumulative_change_rate = 0.0 output: torch.Tensor = executor(*args, **kwargs) if easycache.has_output_prev_norm(): output_change = (easycache.subsample(output, clone=False) - easycache.output_prev_subsampled).flatten().abs().mean() if easycache.verbose: output_change_rate = output_change / easycache.output_prev_norm easycache.output_change_rates.append(output_change_rate.item()) if easycache.has_relative_transformation_rate(): approx_output_change_rate = (easycache.relative_transformation_rate * input_change) / easycache.output_prev_norm easycache.approx_output_change_rates.append(approx_output_change_rate.item()) if easycache.verbose: logging.info(f"LazyCache [verbose] - approx_output_change_rate: {approx_output_change_rate}") if input_change is not None: easycache.relative_transformation_rate = output_change / input_change if easycache.verbose: logging.info(f"LazyCache [verbose] - output_change_rate: {output_change_rate}") # TODO: allow cache_diff to be offloaded easycache.update_cache_diff(output, next_x_prev) easycache.x_prev_subsampled = easycache.subsample(next_x_prev) easycache.output_prev_subsampled = easycache.subsample(output) easycache.output_prev_norm = output.flatten().abs().mean() if easycache.verbose: logging.info(f"LazyCache [verbose] - x_prev_subsampled: {easycache.x_prev_subsampled.shape}") return output def easycache_calc_cond_batch_wrapper(executor, *args, **kwargs): model_options = args[-1] easycache: EasyCacheHolder = model_options["transformer_options"]["easycache"] easycache.skip_current_step = False # TODO: check if first_cond_uuid is active at this timestep; otherwise, EasyCache needs to be partially reset return executor(*args, **kwargs) def easycache_sample_wrapper(executor, *args, **kwargs): """ This OUTER_SAMPLE wrapper makes sure easycache is prepped for current run, and all memory usage is cleared at the end. """ try: guider = executor.class_obj orig_model_options = guider.model_options guider.model_options = comfy.model_patcher.create_model_options_clone(orig_model_options) # clone and prepare timesteps guider.model_options["transformer_options"]["easycache"] = guider.model_options["transformer_options"]["easycache"].clone().prepare_timesteps(guider.model_patcher.model.model_sampling) easycache: Union[EasyCacheHolder, LazyCacheHolder] = guider.model_options['transformer_options']['easycache'] logging.info(f"{easycache.name} enabled - threshold: {easycache.reuse_threshold}, start_percent: {easycache.start_percent}, end_percent: {easycache.end_percent}") return executor(*args, **kwargs) finally: easycache = guider.model_options['transformer_options']['easycache'] output_change_rates = easycache.output_change_rates approx_output_change_rates = easycache.approx_output_change_rates if easycache.verbose: logging.info(f"{easycache.name} [verbose] - output_change_rates {len(output_change_rates)}: {output_change_rates}") logging.info(f"{easycache.name} [verbose] - approx_output_change_rates {len(approx_output_change_rates)}: {approx_output_change_rates}") total_steps = len(args[3])-1 # catch division by zero for log statement; sucks to crash after all sampling is done try: speedup = total_steps/(total_steps-easycache.total_steps_skipped) except ZeroDivisionError: speedup = 1.0 logging.info(f"{easycache.name} - skipped {easycache.total_steps_skipped}/{total_steps} steps ({speedup:.2f}x speedup).") easycache.reset() guider.model_options = orig_model_options class EasyCacheHolder: def __init__(self, reuse_threshold: float, start_percent: float, end_percent: float, subsample_factor: int, offload_cache_diff: bool, verbose: bool=False, output_channels: int=None): self.name = "EasyCache" self.reuse_threshold = reuse_threshold self.start_percent = start_percent self.end_percent = end_percent self.subsample_factor = subsample_factor self.offload_cache_diff = offload_cache_diff self.verbose = verbose # timestep values self.start_t = 0.0 self.end_t = 0.0 # control values self.relative_transformation_rate: float = None self.cumulative_change_rate = 0.0 self.initial_step = True self.skip_current_step = False # cache values self.first_cond_uuid = None self.x_prev_subsampled: torch.Tensor = None self.output_prev_subsampled: torch.Tensor = None self.output_prev_norm: torch.Tensor = None self.uuid_cache_diffs: dict[UUID, torch.Tensor] = {} self.uuid_cache_diffs_audio: dict[UUID, torch.Tensor] = {} self.output_change_rates = [] self.approx_output_change_rates = [] self.total_steps_skipped = 0 # how to deal with mismatched dims self.allow_mismatch = True self.cut_from_start = True self.state_metadata = None self.output_channels = output_channels def is_past_end_timestep(self, timestep: float) -> bool: return not (timestep[0] > self.end_t).item() def should_do_easycache(self, timestep: float) -> bool: return (timestep[0] <= self.start_t).item() def has_x_prev_subsampled(self) -> bool: return self.x_prev_subsampled is not None def has_output_prev_subsampled(self) -> bool: return self.output_prev_subsampled is not None def has_output_prev_norm(self) -> bool: return self.output_prev_norm is not None def has_relative_transformation_rate(self) -> bool: return self.relative_transformation_rate is not None def prepare_timesteps(self, model_sampling): self.start_t = model_sampling.percent_to_sigma(self.start_percent) self.end_t = model_sampling.percent_to_sigma(self.end_percent) return self def subsample(self, x: torch.Tensor, uuids: list[UUID], clone: bool = True) -> torch.Tensor: batch_offset = x.shape[0] // len(uuids) uuid_idx = uuids.index(self.first_cond_uuid) if self.subsample_factor > 1: to_return = x[uuid_idx*batch_offset:(uuid_idx+1)*batch_offset, ..., ::self.subsample_factor, ::self.subsample_factor] if clone: return to_return.clone() return to_return to_return = x[uuid_idx*batch_offset:(uuid_idx+1)*batch_offset, ...] if clone: return to_return.clone() return to_return def can_apply_cache_diff(self, uuids: list[UUID]) -> bool: return all(uuid in self.uuid_cache_diffs for uuid in uuids) def apply_cache_diff(self, x: torch.Tensor, uuids: list[UUID], is_audio: bool = False): if self.first_cond_uuid in uuids and not is_audio: self.total_steps_skipped += 1 cache_diffs = self.uuid_cache_diffs_audio if is_audio else self.uuid_cache_diffs batch_offset = x.shape[0] // len(uuids) for i, uuid in enumerate(uuids): # slice out only what is relevant to this cond batch_slice = [slice(i*batch_offset,(i+1)*batch_offset)] # if cached dims don't match x dims, cut off excess and hope for the best (cosmos world2video) if x.shape[1:] != cache_diffs[uuid].shape[1:]: if not self.allow_mismatch: raise ValueError(f"Cached dims {self.uuid_cache_diffs[uuid].shape} don't match x dims {x.shape} - this is no good") slicing = [] skip_this_dim = True for dim_u, dim_x in zip(cache_diffs[uuid].shape, x.shape): if skip_this_dim: skip_this_dim = False continue if dim_u != dim_x: if self.cut_from_start: slicing.append(slice(dim_x-dim_u, None)) else: slicing.append(slice(None, dim_u)) else: slicing.append(slice(None)) batch_slice = batch_slice + slicing x[tuple(batch_slice)] += cache_diffs[uuid].to(x.device) return x def update_cache_diff(self, output: torch.Tensor, x: torch.Tensor, uuids: list[UUID], is_audio: bool = False): cache_diffs = self.uuid_cache_diffs_audio if is_audio else self.uuid_cache_diffs # if output dims don't match x dims, cut off excess and hope for the best (cosmos world2video) if output.shape[1:] != x.shape[1:]: if not self.allow_mismatch: raise ValueError(f"Output dims {output.shape} don't match x dims {x.shape} - this is no good") slicing = [] skip_dim = True for dim_o, dim_x in zip(output.shape, x.shape): if not skip_dim and dim_o != dim_x: if self.cut_from_start: slicing.append(slice(dim_x-dim_o, None)) else: slicing.append(slice(None, dim_o)) else: slicing.append(slice(None)) skip_dim = False x = x[tuple(slicing)] diff = output - x batch_offset = diff.shape[0] // len(uuids) for i, uuid in enumerate(uuids): cache_diffs[uuid] = diff[i*batch_offset:(i+1)*batch_offset, ...] def has_first_cond_uuid(self, uuids: list[UUID]) -> bool: return self.first_cond_uuid in uuids def check_metadata(self, x: torch.Tensor) -> bool: metadata = (x.device, x.dtype, x.shape[1:]) if self.state_metadata is None: self.state_metadata = metadata return True if metadata == self.state_metadata: return True logging.warn(f"{self.name} - Tensor shape, dtype or device changed, resetting state") self.reset() return False def reset(self): self.relative_transformation_rate = 0.0 self.cumulative_change_rate = 0.0 self.initial_step = True self.skip_current_step = False self.output_change_rates = [] self.first_cond_uuid = None del self.x_prev_subsampled self.x_prev_subsampled = None del self.output_prev_subsampled self.output_prev_subsampled = None del self.output_prev_norm self.output_prev_norm = None del self.uuid_cache_diffs self.uuid_cache_diffs = {} del self.uuid_cache_diffs_audio self.uuid_cache_diffs_audio = {} self.total_steps_skipped = 0 self.state_metadata = None return self def clone(self): return EasyCacheHolder(self.reuse_threshold, self.start_percent, self.end_percent, self.subsample_factor, self.offload_cache_diff, self.verbose, output_channels=self.output_channels) class EasyCacheNode(io.ComfyNode): @classmethod def define_schema(cls) -> io.Schema: return io.Schema( node_id="EasyCache", display_name="EasyCache", description="Native EasyCache implementation.", category="advanced/debug/model", is_experimental=True, inputs=[ io.Model.Input("model", tooltip="The model to add EasyCache to."), io.Float.Input("reuse_threshold", min=0.0, default=0.2, max=3.0, step=0.01, tooltip="The threshold for reusing cached steps.", advanced=True), io.Float.Input("start_percent", min=0.0, default=0.15, max=1.0, step=0.01, tooltip="The relative sampling step to begin use of EasyCache.", advanced=True), io.Float.Input("end_percent", min=0.0, default=0.95, max=1.0, step=0.01, tooltip="The relative sampling step to end use of EasyCache.", advanced=True), io.Boolean.Input("verbose", default=False, tooltip="Whether to log verbose information.", advanced=True), ], outputs=[ io.Model.Output(tooltip="The model with EasyCache."), ], ) @classmethod def execute(cls, model: io.Model.Type, reuse_threshold: float, start_percent: float, end_percent: float, verbose: bool) -> io.NodeOutput: model = model.clone() model.model_options["transformer_options"]["easycache"] = EasyCacheHolder(reuse_threshold, start_percent, end_percent, subsample_factor=8, offload_cache_diff=False, verbose=verbose, output_channels=model.model.latent_format.latent_channels) model.add_wrapper_with_key(comfy.patcher_extension.WrappersMP.OUTER_SAMPLE, "easycache", easycache_sample_wrapper) model.add_wrapper_with_key(comfy.patcher_extension.WrappersMP.CALC_COND_BATCH, "easycache", easycache_calc_cond_batch_wrapper) model.add_wrapper_with_key(comfy.patcher_extension.WrappersMP.DIFFUSION_MODEL, "easycache", easycache_forward_wrapper) return io.NodeOutput(model) class LazyCacheHolder: def __init__(self, reuse_threshold: float, start_percent: float, end_percent: float, subsample_factor: int, offload_cache_diff: bool, verbose: bool=False, output_channels: int=None): self.name = "LazyCache" self.reuse_threshold = reuse_threshold self.start_percent = start_percent self.end_percent = end_percent self.subsample_factor = subsample_factor self.offload_cache_diff = offload_cache_diff self.verbose = verbose # timestep values self.start_t = 0.0 self.end_t = 0.0 # control values self.relative_transformation_rate: float = None self.cumulative_change_rate = 0.0 self.initial_step = True # cache values self.x_prev_subsampled: torch.Tensor = None self.output_prev_subsampled: torch.Tensor = None self.output_prev_norm: torch.Tensor = None self.cache_diff: torch.Tensor = None self.output_change_rates = [] self.approx_output_change_rates = [] self.total_steps_skipped = 0 self.state_metadata = None self.output_channels = output_channels def has_cache_diff(self) -> bool: return self.cache_diff is not None def is_past_end_timestep(self, timestep: float) -> bool: return not (timestep[0] > self.end_t).item() def should_do_easycache(self, timestep: float) -> bool: return (timestep[0] <= self.start_t).item() def has_x_prev_subsampled(self) -> bool: return self.x_prev_subsampled is not None def has_output_prev_subsampled(self) -> bool: return self.output_prev_subsampled is not None def has_output_prev_norm(self) -> bool: return self.output_prev_norm is not None def has_relative_transformation_rate(self) -> bool: return self.relative_transformation_rate is not None def prepare_timesteps(self, model_sampling): self.start_t = model_sampling.percent_to_sigma(self.start_percent) self.end_t = model_sampling.percent_to_sigma(self.end_percent) return self def subsample(self, x: torch.Tensor, clone: bool = True) -> torch.Tensor: if self.subsample_factor > 1: to_return = x[..., ::self.subsample_factor, ::self.subsample_factor] if clone: return to_return.clone() return to_return if clone: return x.clone() return x def apply_cache_diff(self, x: torch.Tensor): self.total_steps_skipped += 1 return x + self.cache_diff.to(x.device) def update_cache_diff(self, output: torch.Tensor, x: torch.Tensor): self.cache_diff = output - x def check_metadata(self, x: torch.Tensor) -> bool: metadata = (x.device, x.dtype, x.shape) if self.state_metadata is None: self.state_metadata = metadata return True if metadata == self.state_metadata: return True logging.warn(f"{self.name} - Tensor shape, dtype or device changed, resetting state") self.reset() return False def reset(self): self.relative_transformation_rate = 0.0 self.cumulative_change_rate = 0.0 self.initial_step = True self.output_change_rates = [] self.approx_output_change_rates = [] del self.cache_diff self.cache_diff = None del self.x_prev_subsampled self.x_prev_subsampled = None del self.output_prev_subsampled self.output_prev_subsampled = None del self.output_prev_norm self.output_prev_norm = None self.total_steps_skipped = 0 self.state_metadata = None return self def clone(self): return LazyCacheHolder(self.reuse_threshold, self.start_percent, self.end_percent, self.subsample_factor, self.offload_cache_diff, self.verbose, output_channels=self.output_channels) class LazyCacheNode(io.ComfyNode): @classmethod def define_schema(cls) -> io.Schema: return io.Schema( node_id="LazyCache", display_name="LazyCache", description="A homebrew version of EasyCache - even 'easier' version of EasyCache to implement. Overall works worse than EasyCache, but better in some rare cases AND universal compatibility with everything in ComfyUI.", category="advanced/debug/model", is_experimental=True, inputs=[ io.Model.Input("model", tooltip="The model to add LazyCache to."), io.Float.Input("reuse_threshold", min=0.0, default=0.2, max=3.0, step=0.01, tooltip="The threshold for reusing cached steps.", advanced=True), io.Float.Input("start_percent", min=0.0, default=0.15, max=1.0, step=0.01, tooltip="The relative sampling step to begin use of LazyCache.", advanced=True), io.Float.Input("end_percent", min=0.0, default=0.95, max=1.0, step=0.01, tooltip="The relative sampling step to end use of LazyCache.", advanced=True), io.Boolean.Input("verbose", default=False, tooltip="Whether to log verbose information.", advanced=True), ], outputs=[ io.Model.Output(tooltip="The model with LazyCache."), ], ) @classmethod def execute(cls, model: io.Model.Type, reuse_threshold: float, start_percent: float, end_percent: float, verbose: bool) -> io.NodeOutput: model = model.clone() model.model_options["transformer_options"]["easycache"] = LazyCacheHolder(reuse_threshold, start_percent, end_percent, subsample_factor=8, offload_cache_diff=False, verbose=verbose, output_channels=model.model.latent_format.latent_channels) model.add_wrapper_with_key(comfy.patcher_extension.WrappersMP.OUTER_SAMPLE, "lazycache", easycache_sample_wrapper) model.add_wrapper_with_key(comfy.patcher_extension.WrappersMP.PREDICT_NOISE, "lazycache", lazycache_predict_noise_wrapper) return io.NodeOutput(model) class EasyCacheExtension(ComfyExtension): async def get_node_list(self) -> list[type[io.ComfyNode]]: return [ EasyCacheNode, LazyCacheNode, ] def comfy_entrypoint(): return EasyCacheExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_extras/nodes_easycache.py", "license": "GNU General Public License v3.0", "lines": 481, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy/ldm/qwen_image/controlnet.py
import torch import math from .model import QwenImageTransformer2DModel from .model import QwenImageTransformerBlock class QwenImageFunControlBlock(QwenImageTransformerBlock): def __init__(self, dim, num_attention_heads, attention_head_dim, has_before_proj=False, dtype=None, device=None, operations=None): super().__init__( dim=dim, num_attention_heads=num_attention_heads, attention_head_dim=attention_head_dim, dtype=dtype, device=device, operations=operations, ) self.has_before_proj = has_before_proj if has_before_proj: self.before_proj = operations.Linear(dim, dim, device=device, dtype=dtype) self.after_proj = operations.Linear(dim, dim, device=device, dtype=dtype) class QwenImageFunControlNetModel(torch.nn.Module): def __init__( self, control_in_features=132, inner_dim=3072, num_attention_heads=24, attention_head_dim=128, num_control_blocks=5, main_model_double=60, injection_layers=(0, 12, 24, 36, 48), dtype=None, device=None, operations=None, ): super().__init__() self.dtype = dtype self.main_model_double = main_model_double self.injection_layers = tuple(injection_layers) # Keep base hint scaling at 1.0 so user-facing strength behaves similarly # to the reference Gen2/VideoX implementation around strength=1. self.hint_scale = 1.0 self.control_img_in = operations.Linear(control_in_features, inner_dim, device=device, dtype=dtype) self.control_blocks = torch.nn.ModuleList([]) for i in range(num_control_blocks): self.control_blocks.append( QwenImageFunControlBlock( dim=inner_dim, num_attention_heads=num_attention_heads, attention_head_dim=attention_head_dim, has_before_proj=(i == 0), dtype=dtype, device=device, operations=operations, ) ) def _process_hint_tokens(self, hint): if hint is None: return None if hint.ndim == 4: hint = hint.unsqueeze(2) # Fun checkpoints are trained with 33 latent channels before 2x2 packing: # [control_latent(16), mask(1), inpaint_latent(16)] -> 132 features. # Default behavior (no inpaint input in stock Apply ControlNet) should use # zeros for mask/inpaint branches, matching VideoX fallback semantics. expected_c = self.control_img_in.weight.shape[1] // 4 if hint.shape[1] == 16 and expected_c == 33: zeros_mask = torch.zeros_like(hint[:, :1]) zeros_inpaint = torch.zeros_like(hint) hint = torch.cat([hint, zeros_mask, zeros_inpaint], dim=1) bs, c, t, h, w = hint.shape hidden_states = torch.nn.functional.pad(hint, (0, w % 2, 0, h % 2)) orig_shape = hidden_states.shape hidden_states = hidden_states.view( orig_shape[0], orig_shape[1], orig_shape[-3], orig_shape[-2] // 2, 2, orig_shape[-1] // 2, 2, ) hidden_states = hidden_states.permute(0, 2, 3, 5, 1, 4, 6) hidden_states = hidden_states.reshape( bs, t * ((h + 1) // 2) * ((w + 1) // 2), c * 4, ) expected_in = self.control_img_in.weight.shape[1] cur_in = hidden_states.shape[-1] if cur_in < expected_in: pad = torch.zeros( (hidden_states.shape[0], hidden_states.shape[1], expected_in - cur_in), device=hidden_states.device, dtype=hidden_states.dtype, ) hidden_states = torch.cat([hidden_states, pad], dim=-1) elif cur_in > expected_in: hidden_states = hidden_states[:, :, :expected_in] return hidden_states def forward( self, x, timesteps, context, attention_mask=None, guidance: torch.Tensor = None, hint=None, transformer_options={}, base_model=None, **kwargs, ): if base_model is None: raise RuntimeError("Qwen Fun ControlNet requires a QwenImage base model at runtime.") encoder_hidden_states_mask = attention_mask # Keep attention mask disabled inside Fun control blocks to mirror # VideoX behavior (they rely on seq lengths for RoPE, not masked attention). encoder_hidden_states_mask = None hidden_states, img_ids, _ = base_model.process_img(x) hint_tokens = self._process_hint_tokens(hint) if hint_tokens is None: raise RuntimeError("Qwen Fun ControlNet requires a control hint image.") if hint_tokens.shape[1] != hidden_states.shape[1]: max_tokens = min(hint_tokens.shape[1], hidden_states.shape[1]) hint_tokens = hint_tokens[:, :max_tokens] hidden_states = hidden_states[:, :max_tokens] img_ids = img_ids[:, :max_tokens] txt_start = round( max( ((x.shape[-1] + (base_model.patch_size // 2)) // base_model.patch_size) // 2, ((x.shape[-2] + (base_model.patch_size // 2)) // base_model.patch_size) // 2, ) ) txt_ids = torch.arange(txt_start, txt_start + context.shape[1], device=x.device).reshape(1, -1, 1).repeat(x.shape[0], 1, 3) ids = torch.cat((txt_ids, img_ids), dim=1) image_rotary_emb = base_model.pe_embedder(ids).to(x.dtype).contiguous() hidden_states = base_model.img_in(hidden_states) encoder_hidden_states = base_model.txt_norm(context) encoder_hidden_states = base_model.txt_in(encoder_hidden_states) if guidance is not None: guidance = guidance * 1000 temb = ( base_model.time_text_embed(timesteps, hidden_states) if guidance is None else base_model.time_text_embed(timesteps, guidance, hidden_states) ) c = self.control_img_in(hint_tokens) for i, block in enumerate(self.control_blocks): if i == 0: c_in = block.before_proj(c) + hidden_states all_c = [] else: all_c = list(torch.unbind(c, dim=0)) c_in = all_c.pop(-1) encoder_hidden_states, c_out = block( hidden_states=c_in, encoder_hidden_states=encoder_hidden_states, encoder_hidden_states_mask=encoder_hidden_states_mask, temb=temb, image_rotary_emb=image_rotary_emb, transformer_options=transformer_options, ) c_skip = block.after_proj(c_out) * self.hint_scale all_c += [c_skip, c_out] c = torch.stack(all_c, dim=0) hints = torch.unbind(c, dim=0)[:-1] controlnet_block_samples = [None] * self.main_model_double for local_idx, base_idx in enumerate(self.injection_layers): if local_idx < len(hints) and base_idx < len(controlnet_block_samples): controlnet_block_samples[base_idx] = hints[local_idx] return {"input": controlnet_block_samples} class QwenImageControlNetModel(QwenImageTransformer2DModel): def __init__( self, extra_condition_channels=0, dtype=None, device=None, operations=None, **kwargs ): super().__init__(final_layer=False, dtype=dtype, device=device, operations=operations, **kwargs) self.main_model_double = 60 # controlnet_blocks self.controlnet_blocks = torch.nn.ModuleList([]) for _ in range(len(self.transformer_blocks)): self.controlnet_blocks.append(operations.Linear(self.inner_dim, self.inner_dim, device=device, dtype=dtype)) self.controlnet_x_embedder = operations.Linear(self.in_channels + extra_condition_channels, self.inner_dim, device=device, dtype=dtype) def forward( self, x, timesteps, context, attention_mask=None, guidance: torch.Tensor = None, ref_latents=None, hint=None, transformer_options={}, **kwargs ): timestep = timesteps encoder_hidden_states = context encoder_hidden_states_mask = attention_mask hidden_states, img_ids, orig_shape = self.process_img(x) hint, _, _ = self.process_img(hint) txt_start = round(max(((x.shape[-1] + (self.patch_size // 2)) // self.patch_size) // 2, ((x.shape[-2] + (self.patch_size // 2)) // self.patch_size) // 2)) txt_ids = torch.arange(txt_start, txt_start + context.shape[1], device=x.device).reshape(1, -1, 1).repeat(x.shape[0], 1, 3) ids = torch.cat((txt_ids, img_ids), dim=1) image_rotary_emb = self.pe_embedder(ids).to(x.dtype).contiguous() del ids, txt_ids, img_ids hidden_states = self.img_in(hidden_states) + self.controlnet_x_embedder(hint) encoder_hidden_states = self.txt_norm(encoder_hidden_states) encoder_hidden_states = self.txt_in(encoder_hidden_states) if guidance is not None: guidance = guidance * 1000 temb = ( self.time_text_embed(timestep, hidden_states) if guidance is None else self.time_text_embed(timestep, guidance, hidden_states) ) repeat = math.ceil(self.main_model_double / len(self.controlnet_blocks)) controlnet_block_samples = () for i, block in enumerate(self.transformer_blocks): encoder_hidden_states, hidden_states = block( hidden_states=hidden_states, encoder_hidden_states=encoder_hidden_states, encoder_hidden_states_mask=encoder_hidden_states_mask, temb=temb, image_rotary_emb=image_rotary_emb, ) controlnet_block_samples = controlnet_block_samples + (self.controlnet_blocks[i](hidden_states),) * repeat return {"input": controlnet_block_samples[:self.main_model_double]}
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/ldm/qwen_image/controlnet.py", "license": "GNU General Public License v3.0", "lines": 228, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_extras/nodes_model_patch.py
import torch from torch import nn import folder_paths import comfy.utils import comfy.ops import comfy.model_management import comfy.ldm.common_dit import comfy.latent_formats import comfy.ldm.lumina.controlnet from comfy.ldm.wan.model_multitalk import WanMultiTalkAttentionBlock, MultiTalkAudioProjModel class BlockWiseControlBlock(torch.nn.Module): # [linear, gelu, linear] def __init__(self, dim: int = 3072, device=None, dtype=None, operations=None): super().__init__() self.x_rms = operations.RMSNorm(dim, eps=1e-6) self.y_rms = operations.RMSNorm(dim, eps=1e-6) self.input_proj = operations.Linear(dim, dim) self.act = torch.nn.GELU() self.output_proj = operations.Linear(dim, dim) def forward(self, x, y): x, y = self.x_rms(x), self.y_rms(y) x = self.input_proj(x + y) x = self.act(x) x = self.output_proj(x) return x class QwenImageBlockWiseControlNet(torch.nn.Module): def __init__( self, num_layers: int = 60, in_dim: int = 64, additional_in_dim: int = 0, dim: int = 3072, device=None, dtype=None, operations=None ): super().__init__() self.additional_in_dim = additional_in_dim self.img_in = operations.Linear(in_dim + additional_in_dim, dim, device=device, dtype=dtype) self.controlnet_blocks = torch.nn.ModuleList( [ BlockWiseControlBlock(dim, device=device, dtype=dtype, operations=operations) for _ in range(num_layers) ] ) def process_input_latent_image(self, latent_image): latent_image[:, :16] = comfy.latent_formats.Wan21().process_in(latent_image[:, :16]) patch_size = 2 hidden_states = comfy.ldm.common_dit.pad_to_patch_size(latent_image, (1, patch_size, patch_size)) orig_shape = hidden_states.shape hidden_states = hidden_states.view(orig_shape[0], orig_shape[1], orig_shape[-2] // 2, 2, orig_shape[-1] // 2, 2) hidden_states = hidden_states.permute(0, 2, 4, 1, 3, 5) hidden_states = hidden_states.reshape(orig_shape[0], (orig_shape[-2] // 2) * (orig_shape[-1] // 2), orig_shape[1] * 4) return self.img_in(hidden_states) def control_block(self, img, controlnet_conditioning, block_id): return self.controlnet_blocks[block_id](img, controlnet_conditioning) class SigLIPMultiFeatProjModel(torch.nn.Module): """ SigLIP Multi-Feature Projection Model for processing style features from different layers and projecting them into a unified hidden space. Args: siglip_token_nums (int): Number of SigLIP tokens, default 257 style_token_nums (int): Number of style tokens, default 256 siglip_token_dims (int): Dimension of SigLIP tokens, default 1536 hidden_size (int): Hidden layer size, default 3072 context_layer_norm (bool): Whether to use context layer normalization, default False """ def __init__( self, siglip_token_nums: int = 729, style_token_nums: int = 64, siglip_token_dims: int = 1152, hidden_size: int = 3072, context_layer_norm: bool = True, device=None, dtype=None, operations=None ): super().__init__() # High-level feature processing (layer -2) self.high_embedding_linear = nn.Sequential( operations.Linear(siglip_token_nums, style_token_nums), nn.SiLU() ) self.high_layer_norm = ( operations.LayerNorm(siglip_token_dims) if context_layer_norm else nn.Identity() ) self.high_projection = operations.Linear(siglip_token_dims, hidden_size, bias=True) # Mid-level feature processing (layer -11) self.mid_embedding_linear = nn.Sequential( operations.Linear(siglip_token_nums, style_token_nums), nn.SiLU() ) self.mid_layer_norm = ( operations.LayerNorm(siglip_token_dims) if context_layer_norm else nn.Identity() ) self.mid_projection = operations.Linear(siglip_token_dims, hidden_size, bias=True) # Low-level feature processing (layer -20) self.low_embedding_linear = nn.Sequential( operations.Linear(siglip_token_nums, style_token_nums), nn.SiLU() ) self.low_layer_norm = ( operations.LayerNorm(siglip_token_dims) if context_layer_norm else nn.Identity() ) self.low_projection = operations.Linear(siglip_token_dims, hidden_size, bias=True) def forward(self, siglip_outputs): """ Forward pass function Args: siglip_outputs: Output from SigLIP model, containing hidden_states Returns: torch.Tensor: Concatenated multi-layer features with shape [bs, 3*style_token_nums, hidden_size] """ dtype = next(self.high_embedding_linear.parameters()).dtype # Process high-level features (layer -2) high_embedding = self._process_layer_features( siglip_outputs[2], self.high_embedding_linear, self.high_layer_norm, self.high_projection, dtype ) # Process mid-level features (layer -11) mid_embedding = self._process_layer_features( siglip_outputs[1], self.mid_embedding_linear, self.mid_layer_norm, self.mid_projection, dtype ) # Process low-level features (layer -20) low_embedding = self._process_layer_features( siglip_outputs[0], self.low_embedding_linear, self.low_layer_norm, self.low_projection, dtype ) # Concatenate features from all layersmodel_patch return torch.cat((high_embedding, mid_embedding, low_embedding), dim=1) def _process_layer_features( self, hidden_states: torch.Tensor, embedding_linear: nn.Module, layer_norm: nn.Module, projection: nn.Module, dtype: torch.dtype ) -> torch.Tensor: """ Helper function to process features from a single layer Args: hidden_states: Input hidden states [bs, seq_len, dim] embedding_linear: Embedding linear layer layer_norm: Layer normalization projection: Projection layer dtype: Target data type Returns: torch.Tensor: Processed features [bs, style_token_nums, hidden_size] """ # Transform dimensions: [bs, seq_len, dim] -> [bs, dim, seq_len] -> [bs, dim, style_token_nums] -> [bs, style_token_nums, dim] embedding = embedding_linear( hidden_states.to(dtype).transpose(1, 2) ).transpose(1, 2) # Apply layer normalization embedding = layer_norm(embedding) # Project to target hidden space embedding = projection(embedding) return embedding def z_image_convert(sd): replace_keys = {".attention.to_out.0.bias": ".attention.out.bias", ".attention.norm_k.weight": ".attention.k_norm.weight", ".attention.norm_q.weight": ".attention.q_norm.weight", ".attention.to_out.0.weight": ".attention.out.weight" } out_sd = {} for k in sorted(sd.keys()): w = sd[k] k_out = k if k_out.endswith(".attention.to_k.weight"): cc = [w] continue if k_out.endswith(".attention.to_q.weight"): cc = [w] + cc continue if k_out.endswith(".attention.to_v.weight"): cc = cc + [w] w = torch.cat(cc, dim=0) k_out = k_out.replace(".attention.to_v.weight", ".attention.qkv.weight") for r, rr in replace_keys.items(): k_out = k_out.replace(r, rr) out_sd[k_out] = w return out_sd class ModelPatchLoader: @classmethod def INPUT_TYPES(s): return {"required": { "name": (folder_paths.get_filename_list("model_patches"), ), }} RETURN_TYPES = ("MODEL_PATCH",) FUNCTION = "load_model_patch" EXPERIMENTAL = True CATEGORY = "advanced/loaders" def load_model_patch(self, name): model_patch_path = folder_paths.get_full_path_or_raise("model_patches", name) sd = comfy.utils.load_torch_file(model_patch_path, safe_load=True) dtype = comfy.utils.weight_dtype(sd) if 'controlnet_blocks.0.y_rms.weight' in sd: additional_in_dim = sd["img_in.weight"].shape[1] - 64 model = QwenImageBlockWiseControlNet(additional_in_dim=additional_in_dim, device=comfy.model_management.unet_offload_device(), dtype=dtype, operations=comfy.ops.manual_cast) elif 'feature_embedder.mid_layer_norm.bias' in sd: sd = comfy.utils.state_dict_prefix_replace(sd, {"feature_embedder.": ""}, filter_keys=True) model = SigLIPMultiFeatProjModel(device=comfy.model_management.unet_offload_device(), dtype=dtype, operations=comfy.ops.manual_cast) elif 'control_all_x_embedder.2-1.weight' in sd: # alipai z image fun controlnet sd = z_image_convert(sd) config = {} if 'control_layers.4.adaLN_modulation.0.weight' not in sd: config['n_control_layers'] = 3 config['additional_in_dim'] = 17 config['refiner_control'] = True if 'control_layers.14.adaLN_modulation.0.weight' in sd: config['n_control_layers'] = 15 config['additional_in_dim'] = 17 config['refiner_control'] = True ref_weight = sd.get("control_noise_refiner.0.after_proj.weight", None) if ref_weight is not None: if torch.count_nonzero(ref_weight) == 0: config['broken'] = True model = comfy.ldm.lumina.controlnet.ZImage_Control(device=comfy.model_management.unet_offload_device(), dtype=dtype, operations=comfy.ops.manual_cast, **config) elif "audio_proj.proj1.weight" in sd: model = MultiTalkModelPatch( audio_window=5, context_tokens=32, vae_scale=4, in_dim=sd["blocks.0.audio_cross_attn.proj.weight"].shape[0], intermediate_dim=sd["audio_proj.proj1.weight"].shape[0], out_dim=sd["audio_proj.norm.weight"].shape[0], device=comfy.model_management.unet_offload_device(), operations=comfy.ops.manual_cast) model_patcher = comfy.model_patcher.CoreModelPatcher(model, load_device=comfy.model_management.get_torch_device(), offload_device=comfy.model_management.unet_offload_device()) model.load_state_dict(sd, assign=model_patcher.is_dynamic()) return (model_patcher,) class DiffSynthCnetPatch: def __init__(self, model_patch, vae, image, strength, mask=None): self.model_patch = model_patch self.vae = vae self.image = image self.strength = strength self.mask = mask self.encoded_image = model_patch.model.process_input_latent_image(self.encode_latent_cond(image)) self.encoded_image_size = (image.shape[1], image.shape[2]) def encode_latent_cond(self, image): latent_image = self.vae.encode(image) if self.model_patch.model.additional_in_dim > 0: if self.mask is None: mask_ = torch.ones_like(latent_image)[:, :self.model_patch.model.additional_in_dim // 4] else: mask_ = comfy.utils.common_upscale(self.mask.mean(dim=1, keepdim=True), latent_image.shape[-1], latent_image.shape[-2], "bilinear", "none") return torch.cat([latent_image, mask_], dim=1) else: return latent_image def __call__(self, kwargs): x = kwargs.get("x") img = kwargs.get("img") block_index = kwargs.get("block_index") spacial_compression = self.vae.spacial_compression_encode() if self.encoded_image is None or self.encoded_image_size != (x.shape[-2] * spacial_compression, x.shape[-1] * spacial_compression): image_scaled = comfy.utils.common_upscale(self.image.movedim(-1, 1), x.shape[-1] * spacial_compression, x.shape[-2] * spacial_compression, "area", "center") loaded_models = comfy.model_management.loaded_models(only_currently_used=True) self.encoded_image = self.model_patch.model.process_input_latent_image(self.encode_latent_cond(image_scaled.movedim(1, -1))) self.encoded_image_size = (image_scaled.shape[-2], image_scaled.shape[-1]) comfy.model_management.load_models_gpu(loaded_models) img[:, :self.encoded_image.shape[1]] += (self.model_patch.model.control_block(img[:, :self.encoded_image.shape[1]], self.encoded_image.to(img.dtype), block_index) * self.strength) kwargs['img'] = img return kwargs def to(self, device_or_dtype): if isinstance(device_or_dtype, torch.device): self.encoded_image = self.encoded_image.to(device_or_dtype) return self def models(self): return [self.model_patch] class ZImageControlPatch: def __init__(self, model_patch, vae, image, strength, inpaint_image=None, mask=None): self.model_patch = model_patch self.vae = vae self.image = image self.inpaint_image = inpaint_image self.mask = mask self.strength = strength self.is_inpaint = self.model_patch.model.additional_in_dim > 0 skip_encoding = False if self.image is not None and self.inpaint_image is not None: if self.image.shape != self.inpaint_image.shape: skip_encoding = True if skip_encoding: self.encoded_image = None else: self.encoded_image = self.encode_latent_cond(self.image, self.inpaint_image) if self.image is None: self.encoded_image_size = (self.inpaint_image.shape[1], self.inpaint_image.shape[2]) else: self.encoded_image_size = (self.image.shape[1], self.image.shape[2]) self.temp_data = None def encode_latent_cond(self, control_image=None, inpaint_image=None): latent_image = None if control_image is not None: latent_image = comfy.latent_formats.Flux().process_in(self.vae.encode(control_image)) if self.is_inpaint: if inpaint_image is None: inpaint_image = torch.ones_like(control_image) * 0.5 if self.mask is not None: mask_inpaint = comfy.utils.common_upscale(self.mask.view(self.mask.shape[0], -1, self.mask.shape[-2], self.mask.shape[-1]).mean(dim=1, keepdim=True), inpaint_image.shape[-2], inpaint_image.shape[-3], "bilinear", "center") inpaint_image = ((inpaint_image - 0.5) * mask_inpaint.movedim(1, -1).round()) + 0.5 inpaint_image_latent = comfy.latent_formats.Flux().process_in(self.vae.encode(inpaint_image)) if self.mask is None: mask_ = torch.zeros_like(inpaint_image_latent)[:, :1] else: mask_ = comfy.utils.common_upscale(self.mask.view(self.mask.shape[0], -1, self.mask.shape[-2], self.mask.shape[-1]).mean(dim=1, keepdim=True).to(device=inpaint_image_latent.device), inpaint_image_latent.shape[-1], inpaint_image_latent.shape[-2], "nearest", "center") if latent_image is None: latent_image = comfy.latent_formats.Flux().process_in(self.vae.encode(torch.ones_like(inpaint_image) * 0.5)) return torch.cat([latent_image, mask_, inpaint_image_latent], dim=1) else: return latent_image def __call__(self, kwargs): x = kwargs.get("x") img = kwargs.get("img") img_input = kwargs.get("img_input") txt = kwargs.get("txt") pe = kwargs.get("pe") vec = kwargs.get("vec") block_index = kwargs.get("block_index") block_type = kwargs.get("block_type", "") spacial_compression = self.vae.spacial_compression_encode() if self.encoded_image is None or self.encoded_image_size != (x.shape[-2] * spacial_compression, x.shape[-1] * spacial_compression): image_scaled = None if self.image is not None: image_scaled = comfy.utils.common_upscale(self.image.movedim(-1, 1), x.shape[-1] * spacial_compression, x.shape[-2] * spacial_compression, "area", "center").movedim(1, -1) self.encoded_image_size = (image_scaled.shape[-3], image_scaled.shape[-2]) inpaint_scaled = None if self.inpaint_image is not None: inpaint_scaled = comfy.utils.common_upscale(self.inpaint_image.movedim(-1, 1), x.shape[-1] * spacial_compression, x.shape[-2] * spacial_compression, "area", "center").movedim(1, -1) self.encoded_image_size = (inpaint_scaled.shape[-3], inpaint_scaled.shape[-2]) loaded_models = comfy.model_management.loaded_models(only_currently_used=True) self.encoded_image = self.encode_latent_cond(image_scaled, inpaint_scaled) comfy.model_management.load_models_gpu(loaded_models) cnet_blocks = self.model_patch.model.n_control_layers div = round(30 / cnet_blocks) cnet_index = (block_index // div) cnet_index_float = (block_index / div) kwargs.pop("img") # we do ops in place kwargs.pop("txt") if cnet_index_float > (cnet_blocks - 1): self.temp_data = None return kwargs if self.temp_data is None or self.temp_data[0] > cnet_index: if block_type == "noise_refiner": self.temp_data = (-3, (None, self.model_patch.model(txt, self.encoded_image.to(img.dtype), pe, vec))) else: self.temp_data = (-1, (None, self.model_patch.model(txt, self.encoded_image.to(img.dtype), pe, vec))) if block_type == "noise_refiner": next_layer = self.temp_data[0] + 1 self.temp_data = (next_layer, self.model_patch.model.forward_noise_refiner_block(block_index, self.temp_data[1][1], img_input[:, :self.temp_data[1][1].shape[1]], None, pe, vec)) if self.temp_data[1][0] is not None: img[:, :self.temp_data[1][0].shape[1]] += (self.temp_data[1][0] * self.strength) else: while self.temp_data[0] < cnet_index and (self.temp_data[0] + 1) < cnet_blocks: next_layer = self.temp_data[0] + 1 self.temp_data = (next_layer, self.model_patch.model.forward_control_block(next_layer, self.temp_data[1][1], img_input[:, :self.temp_data[1][1].shape[1]], None, pe, vec)) if cnet_index_float == self.temp_data[0]: img[:, :self.temp_data[1][0].shape[1]] += (self.temp_data[1][0] * self.strength) if cnet_blocks == self.temp_data[0] + 1: self.temp_data = None return kwargs def to(self, device_or_dtype): if isinstance(device_or_dtype, torch.device): if self.encoded_image is not None: self.encoded_image = self.encoded_image.to(device_or_dtype) self.temp_data = None return self def models(self): return [self.model_patch] class QwenImageDiffsynthControlnet: @classmethod def INPUT_TYPES(s): return {"required": { "model": ("MODEL",), "model_patch": ("MODEL_PATCH",), "vae": ("VAE",), "image": ("IMAGE",), "strength": ("FLOAT", {"default": 1.0, "min": -10.0, "max": 10.0, "step": 0.01}), }, "optional": {"mask": ("MASK",)}} RETURN_TYPES = ("MODEL",) FUNCTION = "diffsynth_controlnet" EXPERIMENTAL = True CATEGORY = "advanced/loaders/qwen" def diffsynth_controlnet(self, model, model_patch, vae, image=None, strength=1.0, inpaint_image=None, mask=None): model_patched = model.clone() if image is not None: image = image[:, :, :, :3] if inpaint_image is not None: inpaint_image = inpaint_image[:, :, :, :3] if mask is not None: if mask.ndim == 3: mask = mask.unsqueeze(1) if mask.ndim == 4: mask = mask.unsqueeze(2) mask = 1.0 - mask if isinstance(model_patch.model, comfy.ldm.lumina.controlnet.ZImage_Control): patch = ZImageControlPatch(model_patch, vae, image, strength, inpaint_image=inpaint_image, mask=mask) model_patched.set_model_noise_refiner_patch(patch) model_patched.set_model_double_block_patch(patch) else: model_patched.set_model_double_block_patch(DiffSynthCnetPatch(model_patch, vae, image, strength, mask)) return (model_patched,) class ZImageFunControlnet(QwenImageDiffsynthControlnet): @classmethod def INPUT_TYPES(s): return {"required": { "model": ("MODEL",), "model_patch": ("MODEL_PATCH",), "vae": ("VAE",), "strength": ("FLOAT", {"default": 1.0, "min": -10.0, "max": 10.0, "step": 0.01}), }, "optional": {"image": ("IMAGE",), "inpaint_image": ("IMAGE",), "mask": ("MASK",)}} CATEGORY = "advanced/loaders/zimage" class UsoStyleProjectorPatch: def __init__(self, model_patch, encoded_image): self.model_patch = model_patch self.encoded_image = encoded_image def __call__(self, kwargs): txt_ids = kwargs.get("txt_ids") txt = kwargs.get("txt") siglip_embedding = self.model_patch.model(self.encoded_image.to(txt.dtype)).to(txt.dtype) txt = torch.cat([siglip_embedding, txt], dim=1) kwargs['txt'] = txt kwargs['txt_ids'] = torch.cat([torch.zeros(siglip_embedding.shape[0], siglip_embedding.shape[1], 3, dtype=txt_ids.dtype, device=txt_ids.device), txt_ids], dim=1) return kwargs def to(self, device_or_dtype): if isinstance(device_or_dtype, torch.device): self.encoded_image = self.encoded_image.to(device_or_dtype) return self def models(self): return [self.model_patch] class USOStyleReference: @classmethod def INPUT_TYPES(s): return {"required": {"model": ("MODEL",), "model_patch": ("MODEL_PATCH",), "clip_vision_output": ("CLIP_VISION_OUTPUT", ), }} RETURN_TYPES = ("MODEL",) FUNCTION = "apply_patch" EXPERIMENTAL = True CATEGORY = "advanced/model_patches/flux" def apply_patch(self, model, model_patch, clip_vision_output): encoded_image = torch.stack((clip_vision_output.all_hidden_states[:, -20], clip_vision_output.all_hidden_states[:, -11], clip_vision_output.penultimate_hidden_states)) model_patched = model.clone() model_patched.set_model_post_input_patch(UsoStyleProjectorPatch(model_patch, encoded_image)) return (model_patched,) class MultiTalkModelPatch(torch.nn.Module): def __init__( self, audio_window: int = 5, intermediate_dim: int = 512, in_dim: int = 5120, out_dim: int = 768, context_tokens: int = 32, vae_scale: int = 4, num_layers: int = 40, device=None, dtype=None, operations=None ): super().__init__() self.audio_proj = MultiTalkAudioProjModel( seq_len=audio_window, seq_len_vf=audio_window+vae_scale-1, intermediate_dim=intermediate_dim, out_dim=out_dim, context_tokens=context_tokens, device=device, dtype=dtype, operations=operations ) self.blocks = torch.nn.ModuleList( [ WanMultiTalkAttentionBlock(in_dim, out_dim, device=device, dtype=dtype, operations=operations) for _ in range(num_layers) ] ) NODE_CLASS_MAPPINGS = { "ModelPatchLoader": ModelPatchLoader, "QwenImageDiffsynthControlnet": QwenImageDiffsynthControlnet, "ZImageFunControlnet": ZImageFunControlnet, "USOStyleReference": USOStyleReference, }
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_extras/nodes_model_patch.py", "license": "GNU General Public License v3.0", "lines": 486, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_api_nodes/nodes_vidu.py
from typing_extensions import override from comfy_api.latest import IO, ComfyExtension, Input from comfy_api_nodes.apis.vidu import ( FrameSetting, SubjectReference, TaskCreationRequest, TaskCreationResponse, TaskExtendCreationRequest, TaskMultiFrameCreationRequest, TaskResult, TaskStatusResponse, ) from comfy_api_nodes.util import ( ApiEndpoint, download_url_to_video_output, get_number_of_images, poll_op, sync_op, upload_image_to_comfyapi, upload_images_to_comfyapi, upload_video_to_comfyapi, validate_image_aspect_ratio, validate_image_dimensions, validate_images_aspect_ratio_closeness, validate_string, validate_video_duration, ) VIDU_TEXT_TO_VIDEO = "/proxy/vidu/text2video" VIDU_IMAGE_TO_VIDEO = "/proxy/vidu/img2video" VIDU_REFERENCE_VIDEO = "/proxy/vidu/reference2video" VIDU_START_END_VIDEO = "/proxy/vidu/start-end2video" VIDU_GET_GENERATION_STATUS = "/proxy/vidu/tasks/%s/creations" async def execute_task( cls: type[IO.ComfyNode], vidu_endpoint: str, payload: TaskCreationRequest | TaskExtendCreationRequest | TaskMultiFrameCreationRequest, max_poll_attempts: int = 320, ) -> list[TaskResult]: task_creation_response = await sync_op( cls, endpoint=ApiEndpoint(path=vidu_endpoint, method="POST"), response_model=TaskCreationResponse, data=payload, ) if task_creation_response.state == "failed": raise RuntimeError(f"Vidu request failed. Code: {task_creation_response.code}") response = await poll_op( cls, ApiEndpoint(path=VIDU_GET_GENERATION_STATUS % task_creation_response.task_id), response_model=TaskStatusResponse, status_extractor=lambda r: r.state, progress_extractor=lambda r: r.progress, price_extractor=lambda r: r.credits * 0.005 if r.credits is not None else None, max_poll_attempts=max_poll_attempts, ) if not response.creations: raise RuntimeError( f"Vidu request does not contain results. State: {response.state}, Error Code: {response.err_code}" ) return response.creations class ViduTextToVideoNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="ViduTextToVideoNode", display_name="Vidu Text To Video Generation", category="api node/video/Vidu", description="Generate video from a text prompt", inputs=[ IO.Combo.Input("model", options=["viduq1"], tooltip="Model name"), IO.String.Input( "prompt", multiline=True, tooltip="A textual description for video generation", ), IO.Int.Input( "duration", default=5, min=5, max=5, step=1, display_mode=IO.NumberDisplay.number, tooltip="Duration of the output video in seconds", optional=True, ), IO.Int.Input( "seed", default=0, min=0, max=2147483647, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, tooltip="Seed for video generation (0 for random)", optional=True, ), IO.Combo.Input( "aspect_ratio", options=["16:9", "9:16", "1:1"], tooltip="The aspect ratio of the output video", optional=True, ), IO.Combo.Input( "resolution", options=["1080p"], tooltip="Supported values may vary by model & duration", optional=True, advanced=True, ), IO.Combo.Input( "movement_amplitude", options=["auto", "small", "medium", "large"], tooltip="The movement amplitude of objects in the frame", optional=True, advanced=True, ), ], outputs=[ IO.Video.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.4}""", ), ) @classmethod async def execute( cls, model: str, prompt: str, duration: int, seed: int, aspect_ratio: str, resolution: str, movement_amplitude: str, ) -> IO.NodeOutput: if not prompt: raise ValueError("The prompt field is required and cannot be empty.") payload = TaskCreationRequest( model=model, prompt=prompt, duration=duration, seed=seed, aspect_ratio=aspect_ratio, resolution=resolution, movement_amplitude=movement_amplitude, ) results = await execute_task(cls, VIDU_TEXT_TO_VIDEO, payload) return IO.NodeOutput(await download_url_to_video_output(results[0].url)) class ViduImageToVideoNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="ViduImageToVideoNode", display_name="Vidu Image To Video Generation", category="api node/video/Vidu", description="Generate video from image and optional prompt", inputs=[ IO.Combo.Input("model", options=["viduq1"], tooltip="Model name"), IO.Image.Input( "image", tooltip="An image to be used as the start frame of the generated video", ), IO.String.Input( "prompt", multiline=True, default="", tooltip="A textual description for video generation", optional=True, ), IO.Int.Input( "duration", default=5, min=5, max=5, step=1, display_mode=IO.NumberDisplay.number, tooltip="Duration of the output video in seconds", optional=True, ), IO.Int.Input( "seed", default=0, min=0, max=2147483647, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, tooltip="Seed for video generation (0 for random)", optional=True, ), IO.Combo.Input( "resolution", options=["1080p"], tooltip="Supported values may vary by model & duration", optional=True, advanced=True, ), IO.Combo.Input( "movement_amplitude", options=["auto", "small", "medium", "large"], tooltip="The movement amplitude of objects in the frame", optional=True, advanced=True, ), ], outputs=[ IO.Video.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.4}""", ), ) @classmethod async def execute( cls, model: str, image: Input.Image, prompt: str, duration: int, seed: int, resolution: str, movement_amplitude: str, ) -> IO.NodeOutput: if get_number_of_images(image) > 1: raise ValueError("Only one input image is allowed.") validate_image_aspect_ratio(image, (1, 4), (4, 1)) payload = TaskCreationRequest( model=model, prompt=prompt, duration=duration, seed=seed, resolution=resolution, movement_amplitude=movement_amplitude, ) payload.images = await upload_images_to_comfyapi( cls, image, max_images=1, mime_type="image/png", ) results = await execute_task(cls, VIDU_IMAGE_TO_VIDEO, payload) return IO.NodeOutput(await download_url_to_video_output(results[0].url)) class ViduReferenceVideoNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="ViduReferenceVideoNode", display_name="Vidu Reference To Video Generation", category="api node/video/Vidu", description="Generate video from multiple images and a prompt", inputs=[ IO.Combo.Input("model", options=["viduq1"], tooltip="Model name"), IO.Image.Input( "images", tooltip="Images to use as references to generate a video with consistent subjects (max 7 images).", ), IO.String.Input( "prompt", multiline=True, tooltip="A textual description for video generation", ), IO.Int.Input( "duration", default=5, min=5, max=5, step=1, display_mode=IO.NumberDisplay.number, tooltip="Duration of the output video in seconds", optional=True, ), IO.Int.Input( "seed", default=0, min=0, max=2147483647, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, tooltip="Seed for video generation (0 for random)", optional=True, ), IO.Combo.Input( "aspect_ratio", options=["16:9", "9:16", "1:1"], tooltip="The aspect ratio of the output video", optional=True, ), IO.Combo.Input( "resolution", options=["1080p"], tooltip="Supported values may vary by model & duration", optional=True, advanced=True, ), IO.Combo.Input( "movement_amplitude", options=["auto", "small", "medium", "large"], tooltip="The movement amplitude of objects in the frame", optional=True, advanced=True, ), ], outputs=[ IO.Video.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.4}""", ), ) @classmethod async def execute( cls, model: str, images: Input.Image, prompt: str, duration: int, seed: int, aspect_ratio: str, resolution: str, movement_amplitude: str, ) -> IO.NodeOutput: if not prompt: raise ValueError("The prompt field is required and cannot be empty.") a = get_number_of_images(images) if a > 7: raise ValueError("Too many images, maximum allowed is 7.") for image in images: validate_image_aspect_ratio(image, (1, 4), (4, 1)) validate_image_dimensions(image, min_width=128, min_height=128) payload = TaskCreationRequest( model=model, prompt=prompt, duration=duration, seed=seed, aspect_ratio=aspect_ratio, resolution=resolution, movement_amplitude=movement_amplitude, ) payload.images = await upload_images_to_comfyapi( cls, images, max_images=7, mime_type="image/png", ) results = await execute_task(cls, VIDU_REFERENCE_VIDEO, payload) return IO.NodeOutput(await download_url_to_video_output(results[0].url)) class ViduStartEndToVideoNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="ViduStartEndToVideoNode", display_name="Vidu Start End To Video Generation", category="api node/video/Vidu", description="Generate a video from start and end frames and a prompt", inputs=[ IO.Combo.Input("model", options=["viduq1"], tooltip="Model name"), IO.Image.Input( "first_frame", tooltip="Start frame", ), IO.Image.Input( "end_frame", tooltip="End frame", ), IO.String.Input( "prompt", multiline=True, tooltip="A textual description for video generation", optional=True, ), IO.Int.Input( "duration", default=5, min=5, max=5, step=1, display_mode=IO.NumberDisplay.number, tooltip="Duration of the output video in seconds", optional=True, ), IO.Int.Input( "seed", default=0, min=0, max=2147483647, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, tooltip="Seed for video generation (0 for random)", optional=True, ), IO.Combo.Input( "resolution", options=["1080p"], tooltip="Supported values may vary by model & duration", optional=True, advanced=True, ), IO.Combo.Input( "movement_amplitude", options=["auto", "small", "medium", "large"], tooltip="The movement amplitude of objects in the frame", optional=True, advanced=True, ), ], outputs=[ IO.Video.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.4}""", ), ) @classmethod async def execute( cls, model: str, first_frame: Input.Image, end_frame: Input.Image, prompt: str, duration: int, seed: int, resolution: str, movement_amplitude: str, ) -> IO.NodeOutput: validate_images_aspect_ratio_closeness(first_frame, end_frame, min_rel=0.8, max_rel=1.25, strict=False) payload = TaskCreationRequest( model=model, prompt=prompt, duration=duration, seed=seed, resolution=resolution, movement_amplitude=movement_amplitude, ) payload.images = [ (await upload_images_to_comfyapi(cls, frame, max_images=1, mime_type="image/png"))[0] for frame in (first_frame, end_frame) ] results = await execute_task(cls, VIDU_START_END_VIDEO, payload) return IO.NodeOutput(await download_url_to_video_output(results[0].url)) class Vidu2TextToVideoNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="Vidu2TextToVideoNode", display_name="Vidu2 Text-to-Video Generation", category="api node/video/Vidu", description="Generate video from a text prompt", inputs=[ IO.Combo.Input("model", options=["viduq2"]), IO.String.Input( "prompt", multiline=True, tooltip="A textual description for video generation, with a maximum length of 2000 characters.", ), IO.Int.Input( "duration", default=5, min=1, max=10, step=1, display_mode=IO.NumberDisplay.slider, ), IO.Int.Input( "seed", default=1, min=0, max=2147483647, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, ), IO.Combo.Input("aspect_ratio", options=["16:9", "9:16", "3:4", "4:3", "1:1"]), IO.Combo.Input("resolution", options=["720p", "1080p"], advanced=True), IO.Boolean.Input( "background_music", default=False, tooltip="Whether to add background music to the generated video.", advanced=True, ), ], outputs=[ IO.Video.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["duration", "resolution"]), expr=""" ( $is1080 := widgets.resolution = "1080p"; $base := $is1080 ? 0.1 : 0.075; $perSec := $is1080 ? 0.05 : 0.025; {"type":"usd","usd": $base + $perSec * (widgets.duration - 1)} ) """, ), ) @classmethod async def execute( cls, model: str, prompt: str, duration: int, seed: int, aspect_ratio: str, resolution: str, background_music: bool, ) -> IO.NodeOutput: validate_string(prompt, min_length=1, max_length=2000) results = await execute_task( cls, VIDU_TEXT_TO_VIDEO, TaskCreationRequest( model=model, prompt=prompt, duration=duration, seed=seed, aspect_ratio=aspect_ratio, resolution=resolution, bgm=background_music, ), ) return IO.NodeOutput(await download_url_to_video_output(results[0].url)) class Vidu2ImageToVideoNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="Vidu2ImageToVideoNode", display_name="Vidu2 Image-to-Video Generation", category="api node/video/Vidu", description="Generate a video from an image and an optional prompt.", inputs=[ IO.Combo.Input("model", options=["viduq2-pro-fast", "viduq2-pro", "viduq2-turbo"]), IO.Image.Input( "image", tooltip="An image to be used as the start frame of the generated video.", ), IO.String.Input( "prompt", multiline=True, default="", tooltip="An optional text prompt for video generation (max 2000 characters).", ), IO.Int.Input( "duration", default=5, min=1, max=10, step=1, display_mode=IO.NumberDisplay.slider, ), IO.Int.Input( "seed", default=1, min=0, max=2147483647, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, ), IO.Combo.Input( "resolution", options=["720p", "1080p"], advanced=True, ), IO.Combo.Input( "movement_amplitude", options=["auto", "small", "medium", "large"], tooltip="The movement amplitude of objects in the frame.", advanced=True, ), ], outputs=[ IO.Video.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["model", "duration", "resolution"]), expr=""" ( $m := widgets.model; $d := widgets.duration; $is1080 := widgets.resolution = "1080p"; $contains($m, "pro-fast") ? ( $base := $is1080 ? 0.08 : 0.04; $perSec := $is1080 ? 0.02 : 0.01; {"type":"usd","usd": $base + $perSec * ($d - 1)} ) : $contains($m, "pro") ? ( $base := $is1080 ? 0.275 : 0.075; $perSec := $is1080 ? 0.075 : 0.05; {"type":"usd","usd": $base + $perSec * ($d - 1)} ) : $contains($m, "turbo") ? ( $is1080 ? {"type":"usd","usd": 0.175 + 0.05 * ($d - 1)} : ( $d <= 1 ? {"type":"usd","usd": 0.04} : $d <= 2 ? {"type":"usd","usd": 0.05} : {"type":"usd","usd": 0.05 + 0.05 * ($d - 2)} ) ) : {"type":"usd","usd": 0.04} ) """, ), ) @classmethod async def execute( cls, model: str, image: Input.Image, prompt: str, duration: int, seed: int, resolution: str, movement_amplitude: str, ) -> IO.NodeOutput: if get_number_of_images(image) > 1: raise ValueError("Only one input image is allowed.") validate_image_aspect_ratio(image, (1, 4), (4, 1)) validate_string(prompt, max_length=2000) results = await execute_task( cls, VIDU_IMAGE_TO_VIDEO, TaskCreationRequest( model=model, prompt=prompt, duration=duration, seed=seed, resolution=resolution, movement_amplitude=movement_amplitude, images=await upload_images_to_comfyapi( cls, image, max_images=1, mime_type="image/png", ), ), ) return IO.NodeOutput(await download_url_to_video_output(results[0].url)) class Vidu2ReferenceVideoNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="Vidu2ReferenceVideoNode", display_name="Vidu2 Reference-to-Video Generation", category="api node/video/Vidu", description="Generate a video from multiple reference images and a prompt.", inputs=[ IO.Combo.Input("model", options=["viduq2"]), IO.Autogrow.Input( "subjects", template=IO.Autogrow.TemplateNames( IO.Image.Input("reference_images"), names=["subject1", "subject2", "subject3", "subject4", "subject5", "subject6", "subject7"], min=1, ), tooltip="For each subject, provide up to 3 reference images (7 images total across all subjects). " "Reference them in prompts via @subject{subject_id}.", ), IO.String.Input( "prompt", multiline=True, tooltip="When enabled, the video will include generated speech and background music " "based on the prompt.", ), IO.Boolean.Input( "audio", default=False, tooltip="When enabled video will contain generated speech and background music based on the prompt.", advanced=True, ), IO.Int.Input( "duration", default=5, min=1, max=10, step=1, display_mode=IO.NumberDisplay.slider, ), IO.Int.Input( "seed", default=1, min=0, max=2147483647, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, ), IO.Combo.Input("aspect_ratio", options=["16:9", "9:16", "4:3", "3:4", "1:1"]), IO.Combo.Input("resolution", options=["720p", "1080p"], advanced=True), IO.Combo.Input( "movement_amplitude", options=["auto", "small", "medium", "large"], tooltip="The movement amplitude of objects in the frame.", advanced=True, ), ], outputs=[ IO.Video.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["audio", "duration", "resolution"]), expr=""" ( $is1080 := widgets.resolution = "1080p"; $base := $is1080 ? 0.375 : 0.125; $perSec := $is1080 ? 0.05 : 0.025; $audioCost := widgets.audio = true ? 0.075 : 0; {"type":"usd","usd": $base + $perSec * (widgets.duration - 1) + $audioCost} ) """, ), ) @classmethod async def execute( cls, model: str, subjects: IO.Autogrow.Type, prompt: str, audio: bool, duration: int, seed: int, aspect_ratio: str, resolution: str, movement_amplitude: str, ) -> IO.NodeOutput: validate_string(prompt, min_length=1, max_length=2000) total_images = 0 for i in subjects: if get_number_of_images(subjects[i]) > 3: raise ValueError("Maximum number of images per subject is 3.") for im in subjects[i]: total_images += 1 validate_image_aspect_ratio(im, (1, 4), (4, 1)) validate_image_dimensions(im, min_width=128, min_height=128) if total_images > 7: raise ValueError("Too many reference images; the maximum allowed is 7.") subjects_param: list[SubjectReference] = [] for i in subjects: subjects_param.append( SubjectReference( id=i, images=await upload_images_to_comfyapi( cls, subjects[i], max_images=3, mime_type="image/png", wait_label=f"Uploading reference images for {i}", ), ), ) payload = TaskCreationRequest( model=model, prompt=prompt, audio=audio, duration=duration, seed=seed, aspect_ratio=aspect_ratio, resolution=resolution, movement_amplitude=movement_amplitude, subjects=subjects_param, ) results = await execute_task(cls, VIDU_REFERENCE_VIDEO, payload) return IO.NodeOutput(await download_url_to_video_output(results[0].url)) class Vidu2StartEndToVideoNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="Vidu2StartEndToVideoNode", display_name="Vidu2 Start/End Frame-to-Video Generation", category="api node/video/Vidu", description="Generate a video from a start frame, an end frame, and a prompt.", inputs=[ IO.Combo.Input("model", options=["viduq2-pro-fast", "viduq2-pro", "viduq2-turbo"]), IO.Image.Input("first_frame"), IO.Image.Input("end_frame"), IO.String.Input( "prompt", multiline=True, tooltip="Prompt description (max 2000 characters).", ), IO.Int.Input( "duration", default=5, min=2, max=8, step=1, display_mode=IO.NumberDisplay.slider, ), IO.Int.Input( "seed", default=1, min=0, max=2147483647, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, ), IO.Combo.Input("resolution", options=["720p", "1080p"], advanced=True), IO.Combo.Input( "movement_amplitude", options=["auto", "small", "medium", "large"], tooltip="The movement amplitude of objects in the frame.", advanced=True, ), ], outputs=[ IO.Video.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["model", "duration", "resolution"]), expr=""" ( $m := widgets.model; $d := widgets.duration; $is1080 := widgets.resolution = "1080p"; $contains($m, "pro-fast") ? ( $base := $is1080 ? 0.08 : 0.04; $perSec := $is1080 ? 0.02 : 0.01; {"type":"usd","usd": $base + $perSec * ($d - 1)} ) : $contains($m, "pro") ? ( $base := $is1080 ? 0.275 : 0.075; $perSec := $is1080 ? 0.075 : 0.05; {"type":"usd","usd": $base + $perSec * ($d - 1)} ) : $contains($m, "turbo") ? ( $is1080 ? {"type":"usd","usd": 0.175 + 0.05 * ($d - 1)} : ( $d <= 2 ? {"type":"usd","usd": 0.05} : {"type":"usd","usd": 0.05 + 0.05 * ($d - 2)} ) ) : {"type":"usd","usd": 0.04} ) """, ), ) @classmethod async def execute( cls, model: str, first_frame: Input.Image, end_frame: Input.Image, prompt: str, duration: int, seed: int, resolution: str, movement_amplitude: str, ) -> IO.NodeOutput: validate_string(prompt, max_length=2000) if get_number_of_images(first_frame) > 1: raise ValueError("Only one input image is allowed for `first_frame`.") if get_number_of_images(end_frame) > 1: raise ValueError("Only one input image is allowed for `end_frame`.") validate_images_aspect_ratio_closeness(first_frame, end_frame, min_rel=0.8, max_rel=1.25, strict=False) payload = TaskCreationRequest( model=model, prompt=prompt, duration=duration, seed=seed, resolution=resolution, movement_amplitude=movement_amplitude, images=[ (await upload_images_to_comfyapi(cls, frame, max_images=1, mime_type="image/png"))[0] for frame in (first_frame, end_frame) ], ) results = await execute_task(cls, VIDU_START_END_VIDEO, payload) return IO.NodeOutput(await download_url_to_video_output(results[0].url)) class ViduExtendVideoNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="ViduExtendVideoNode", display_name="Vidu Video Extension", category="api node/video/Vidu", description="Extend an existing video by generating additional frames.", inputs=[ IO.DynamicCombo.Input( "model", options=[ IO.DynamicCombo.Option( "viduq2-pro", [ IO.Int.Input( "duration", default=4, min=1, max=7, step=1, display_mode=IO.NumberDisplay.slider, tooltip="Duration of the extended video in seconds.", ), IO.Combo.Input( "resolution", options=["720p", "1080p"], tooltip="Resolution of the output video.", ), ], ), IO.DynamicCombo.Option( "viduq2-turbo", [ IO.Int.Input( "duration", default=4, min=1, max=7, step=1, display_mode=IO.NumberDisplay.slider, tooltip="Duration of the extended video in seconds.", ), IO.Combo.Input( "resolution", options=["720p", "1080p"], tooltip="Resolution of the output video.", ), ], ), ], tooltip="Model to use for video extension.", ), IO.Video.Input( "video", tooltip="The source video to extend.", ), IO.String.Input( "prompt", multiline=True, default="", tooltip="An optional text prompt for the extended video (max 2000 characters).", ), IO.Int.Input( "seed", default=1, min=0, max=2147483647, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, ), IO.Image.Input("end_frame", optional=True), ], outputs=[ IO.Video.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["model", "model.duration", "model.resolution"]), expr=""" ( $m := widgets.model; $d := $lookup(widgets, "model.duration"); $res := $lookup(widgets, "model.resolution"); $contains($m, "pro") ? ( $base := $lookup({"720p": 0.15, "1080p": 0.3}, $res); $perSec := $lookup({"720p": 0.05, "1080p": 0.075}, $res); {"type":"usd","usd": $base + $perSec * ($d - 1)} ) : ( $base := $lookup({"720p": 0.075, "1080p": 0.2}, $res); $perSec := $lookup({"720p": 0.025, "1080p": 0.05}, $res); {"type":"usd","usd": $base + $perSec * ($d - 1)} ) ) """, ), ) @classmethod async def execute( cls, model: dict, video: Input.Video, prompt: str, seed: int, end_frame: Input.Image | None = None, ) -> IO.NodeOutput: validate_string(prompt, max_length=2000) validate_video_duration(video, min_duration=4, max_duration=55) image_url = None if end_frame is not None: validate_image_aspect_ratio(end_frame, (1, 4), (4, 1)) validate_image_dimensions(end_frame, min_width=128, min_height=128) image_url = await upload_image_to_comfyapi(cls, end_frame, wait_label="Uploading end frame") results = await execute_task( cls, "/proxy/vidu/extend", TaskExtendCreationRequest( model=model["model"], prompt=prompt, duration=model["duration"], seed=seed, resolution=model["resolution"], video_url=await upload_video_to_comfyapi(cls, video, wait_label="Uploading video"), images=[image_url] if image_url else None, ), max_poll_attempts=480, ) return IO.NodeOutput(await download_url_to_video_output(results[0].url)) def _generate_frame_inputs(count: int) -> list: """Generate input widgets for a given number of frames.""" inputs = [] for i in range(1, count + 1): inputs.extend( [ IO.String.Input( f"prompt{i}", multiline=True, default="", tooltip=f"Text prompt for frame {i} transition.", ), IO.Image.Input( f"end_image{i}", tooltip=f"End frame image for segment {i}. Aspect ratio must be between 1:4 and 4:1.", ), IO.Int.Input( f"duration{i}", default=4, min=2, max=7, step=1, display_mode=IO.NumberDisplay.slider, tooltip=f"Duration for segment {i} in seconds.", ), ] ) return inputs class ViduMultiFrameVideoNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="ViduMultiFrameVideoNode", display_name="Vidu Multi-Frame Video Generation", category="api node/video/Vidu", description="Generate a video with multiple keyframe transitions.", inputs=[ IO.Combo.Input("model", options=["viduq2-pro", "viduq2-turbo"]), IO.Image.Input( "start_image", tooltip="The starting frame image. Aspect ratio must be between 1:4 and 4:1.", ), IO.Int.Input( "seed", default=1, min=0, max=2147483647, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, ), IO.Combo.Input("resolution", options=["720p", "1080p"]), IO.DynamicCombo.Input( "frames", options=[ IO.DynamicCombo.Option("2", _generate_frame_inputs(2)), IO.DynamicCombo.Option("3", _generate_frame_inputs(3)), IO.DynamicCombo.Option("4", _generate_frame_inputs(4)), IO.DynamicCombo.Option("5", _generate_frame_inputs(5)), IO.DynamicCombo.Option("6", _generate_frame_inputs(6)), IO.DynamicCombo.Option("7", _generate_frame_inputs(7)), IO.DynamicCombo.Option("8", _generate_frame_inputs(8)), IO.DynamicCombo.Option("9", _generate_frame_inputs(9)), ], tooltip="Number of keyframe transitions (2-9).", ), ], outputs=[ IO.Video.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends( widgets=[ "model", "resolution", "frames", "frames.duration1", "frames.duration2", "frames.duration3", "frames.duration4", "frames.duration5", "frames.duration6", "frames.duration7", "frames.duration8", "frames.duration9", ] ), expr=""" ( $m := widgets.model; $n := $number(widgets.frames); $is1080 := widgets.resolution = "1080p"; $d1 := $lookup(widgets, "frames.duration1"); $d2 := $lookup(widgets, "frames.duration2"); $d3 := $n >= 3 ? $lookup(widgets, "frames.duration3") : 0; $d4 := $n >= 4 ? $lookup(widgets, "frames.duration4") : 0; $d5 := $n >= 5 ? $lookup(widgets, "frames.duration5") : 0; $d6 := $n >= 6 ? $lookup(widgets, "frames.duration6") : 0; $d7 := $n >= 7 ? $lookup(widgets, "frames.duration7") : 0; $d8 := $n >= 8 ? $lookup(widgets, "frames.duration8") : 0; $d9 := $n >= 9 ? $lookup(widgets, "frames.duration9") : 0; $totalDuration := $d1 + $d2 + $d3 + $d4 + $d5 + $d6 + $d7 + $d8 + $d9; $contains($m, "pro") ? ( $base := $is1080 ? 0.3 : 0.15; $perSec := $is1080 ? 0.075 : 0.05; {"type":"usd","usd": $n * $base + $perSec * $totalDuration} ) : ( $base := $is1080 ? 0.2 : 0.075; $perSec := $is1080 ? 0.05 : 0.025; {"type":"usd","usd": $n * $base + $perSec * $totalDuration} ) ) """, ), ) @classmethod async def execute( cls, model: str, start_image: Input.Image, seed: int, resolution: str, frames: dict, ) -> IO.NodeOutput: validate_image_aspect_ratio(start_image, (1, 4), (4, 1)) frame_count = int(frames["frames"]) image_settings: list[FrameSetting] = [] for i in range(1, frame_count + 1): validate_image_aspect_ratio(frames[f"end_image{i}"], (1, 4), (4, 1)) validate_string(frames[f"prompt{i}"], max_length=2000) start_image_url = await upload_image_to_comfyapi( cls, start_image, mime_type="image/png", wait_label="Uploading start image", ) for i in range(1, frame_count + 1): image_settings.append( FrameSetting( prompt=frames[f"prompt{i}"], key_image=await upload_image_to_comfyapi( cls, frames[f"end_image{i}"], mime_type="image/png", wait_label=f"Uploading end image({i})", ), duration=frames[f"duration{i}"], ) ) results = await execute_task( cls, "/proxy/vidu/multiframe", TaskMultiFrameCreationRequest( model=model, seed=seed, resolution=resolution, start_image=start_image_url, image_settings=image_settings, ), max_poll_attempts=480 * frame_count, ) return IO.NodeOutput(await download_url_to_video_output(results[0].url)) class Vidu3TextToVideoNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="Vidu3TextToVideoNode", display_name="Vidu Q3 Text-to-Video Generation", category="api node/video/Vidu", description="Generate video from a text prompt.", inputs=[ IO.DynamicCombo.Input( "model", options=[ IO.DynamicCombo.Option( "viduq3-pro", [ IO.Combo.Input( "aspect_ratio", options=["16:9", "9:16", "3:4", "4:3", "1:1"], tooltip="The aspect ratio of the output video.", ), IO.Combo.Input( "resolution", options=["720p", "1080p"], tooltip="Resolution of the output video.", ), IO.Int.Input( "duration", default=5, min=1, max=16, step=1, display_mode=IO.NumberDisplay.slider, tooltip="Duration of the output video in seconds.", ), IO.Boolean.Input( "audio", default=False, tooltip="When enabled, outputs video with sound " "(including dialogue and sound effects).", ), ], ), IO.DynamicCombo.Option( "viduq3-turbo", [ IO.Combo.Input( "aspect_ratio", options=["16:9", "9:16", "3:4", "4:3", "1:1"], tooltip="The aspect ratio of the output video.", ), IO.Combo.Input( "resolution", options=["720p", "1080p"], tooltip="Resolution of the output video.", ), IO.Int.Input( "duration", default=5, min=1, max=16, step=1, display_mode=IO.NumberDisplay.slider, tooltip="Duration of the output video in seconds.", ), IO.Boolean.Input( "audio", default=False, tooltip="When enabled, outputs video with sound " "(including dialogue and sound effects).", ), ], ), ], tooltip="Model to use for video generation.", ), IO.String.Input( "prompt", multiline=True, tooltip="A textual description for video generation, with a maximum length of 2000 characters.", ), IO.Int.Input( "seed", default=1, min=0, max=2147483647, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, ), ], outputs=[ IO.Video.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["model", "model.duration", "model.resolution"]), expr=""" ( $res := $lookup(widgets, "model.resolution"); $d := $lookup(widgets, "model.duration"); $contains(widgets.model, "turbo") ? ( $rate := $lookup({"720p": 0.06, "1080p": 0.08}, $res); {"type":"usd","usd": $rate * $d} ) : ( $rate := $lookup({"720p": 0.15, "1080p": 0.16}, $res); {"type":"usd","usd": $rate * $d} ) ) """, ), ) @classmethod async def execute( cls, model: dict, prompt: str, seed: int, ) -> IO.NodeOutput: validate_string(prompt, min_length=1, max_length=2000) results = await execute_task( cls, VIDU_TEXT_TO_VIDEO, TaskCreationRequest( model=model["model"], prompt=prompt, duration=model["duration"], seed=seed, aspect_ratio=model["aspect_ratio"], resolution=model["resolution"], audio=model["audio"], ), max_poll_attempts=640, ) return IO.NodeOutput(await download_url_to_video_output(results[0].url)) class Vidu3ImageToVideoNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="Vidu3ImageToVideoNode", display_name="Vidu Q3 Image-to-Video Generation", category="api node/video/Vidu", description="Generate a video from an image and an optional prompt.", inputs=[ IO.DynamicCombo.Input( "model", options=[ IO.DynamicCombo.Option( "viduq3-pro", [ IO.Combo.Input( "resolution", options=["720p", "1080p", "2K"], tooltip="Resolution of the output video.", ), IO.Int.Input( "duration", default=5, min=1, max=16, step=1, display_mode=IO.NumberDisplay.slider, tooltip="Duration of the output video in seconds.", ), IO.Boolean.Input( "audio", default=False, tooltip="When enabled, outputs video with sound " "(including dialogue and sound effects).", ), ], ), IO.DynamicCombo.Option( "viduq3-turbo", [ IO.Combo.Input( "resolution", options=["720p", "1080p"], tooltip="Resolution of the output video.", ), IO.Int.Input( "duration", default=5, min=1, max=16, step=1, display_mode=IO.NumberDisplay.slider, tooltip="Duration of the output video in seconds.", ), IO.Boolean.Input( "audio", default=False, tooltip="When enabled, outputs video with sound " "(including dialogue and sound effects).", ), ], ), ], tooltip="Model to use for video generation.", ), IO.Image.Input( "image", tooltip="An image to be used as the start frame of the generated video.", ), IO.String.Input( "prompt", multiline=True, default="", tooltip="An optional text prompt for video generation (max 2000 characters).", ), IO.Int.Input( "seed", default=1, min=0, max=2147483647, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, ), ], outputs=[ IO.Video.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["model", "model.duration", "model.resolution"]), expr=""" ( $res := $lookup(widgets, "model.resolution"); $d := $lookup(widgets, "model.duration"); $contains(widgets.model, "turbo") ? ( $rate := $lookup({"720p": 0.06, "1080p": 0.08}, $res); {"type":"usd","usd": $rate * $d} ) : ( $rate := $lookup({"720p": 0.15, "1080p": 0.16, "2k": 0.2}, $res); {"type":"usd","usd": $rate * $d} ) ) """, ), ) @classmethod async def execute( cls, model: dict, image: Input.Image, prompt: str, seed: int, ) -> IO.NodeOutput: validate_image_aspect_ratio(image, (1, 4), (4, 1)) validate_string(prompt, max_length=2000) results = await execute_task( cls, VIDU_IMAGE_TO_VIDEO, TaskCreationRequest( model=model["model"], prompt=prompt, duration=model["duration"], seed=seed, resolution=model["resolution"], audio=model["audio"], images=[await upload_image_to_comfyapi(cls, image)], ), max_poll_attempts=720, ) return IO.NodeOutput(await download_url_to_video_output(results[0].url)) class Vidu3StartEndToVideoNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="Vidu3StartEndToVideoNode", display_name="Vidu Q3 Start/End Frame-to-Video Generation", category="api node/video/Vidu", description="Generate a video from a start frame, an end frame, and a prompt.", inputs=[ IO.DynamicCombo.Input( "model", options=[ IO.DynamicCombo.Option( "viduq3-pro", [ IO.Combo.Input( "resolution", options=["720p", "1080p"], tooltip="Resolution of the output video.", ), IO.Int.Input( "duration", default=5, min=1, max=16, step=1, display_mode=IO.NumberDisplay.slider, tooltip="Duration of the output video in seconds.", ), IO.Boolean.Input( "audio", default=False, tooltip="When enabled, outputs video with sound " "(including dialogue and sound effects).", ), ], ), IO.DynamicCombo.Option( "viduq3-turbo", [ IO.Combo.Input( "resolution", options=["720p", "1080p"], tooltip="Resolution of the output video.", ), IO.Int.Input( "duration", default=5, min=1, max=16, step=1, display_mode=IO.NumberDisplay.slider, tooltip="Duration of the output video in seconds.", ), IO.Boolean.Input( "audio", default=False, tooltip="When enabled, outputs video with sound " "(including dialogue and sound effects).", ), ], ), ], tooltip="Model to use for video generation.", ), IO.Image.Input("first_frame"), IO.Image.Input("end_frame"), IO.String.Input( "prompt", multiline=True, tooltip="Prompt description (max 2000 characters).", ), IO.Int.Input( "seed", default=1, min=0, max=2147483647, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, ), ], outputs=[ IO.Video.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["model", "model.duration", "model.resolution"]), expr=""" ( $res := $lookup(widgets, "model.resolution"); $d := $lookup(widgets, "model.duration"); $contains(widgets.model, "turbo") ? ( $rate := $lookup({"720p": 0.06, "1080p": 0.08}, $res); {"type":"usd","usd": $rate * $d} ) : ( $rate := $lookup({"720p": 0.15, "1080p": 0.16}, $res); {"type":"usd","usd": $rate * $d} ) ) """, ), ) @classmethod async def execute( cls, model: dict, first_frame: Input.Image, end_frame: Input.Image, prompt: str, seed: int, ) -> IO.NodeOutput: validate_string(prompt, max_length=2000) validate_images_aspect_ratio_closeness(first_frame, end_frame, min_rel=0.8, max_rel=1.25, strict=False) payload = TaskCreationRequest( model=model["model"], prompt=prompt, duration=model["duration"], seed=seed, resolution=model["resolution"], audio=model["audio"], images=[ (await upload_images_to_comfyapi(cls, frame, max_images=1, mime_type="image/png"))[0] for frame in (first_frame, end_frame) ], ) results = await execute_task(cls, VIDU_START_END_VIDEO, payload) return IO.NodeOutput(await download_url_to_video_output(results[0].url)) class ViduExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[IO.ComfyNode]]: return [ ViduTextToVideoNode, ViduImageToVideoNode, ViduReferenceVideoNode, ViduStartEndToVideoNode, Vidu2TextToVideoNode, Vidu2ImageToVideoNode, Vidu2ReferenceVideoNode, Vidu2StartEndToVideoNode, ViduExtendVideoNode, ViduMultiFrameVideoNode, Vidu3TextToVideoNode, Vidu3ImageToVideoNode, Vidu3StartEndToVideoNode, ] async def comfy_entrypoint() -> ViduExtension: return ViduExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api_nodes/nodes_vidu.py", "license": "GNU General Public License v3.0", "lines": 1666, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy/text_encoders/qwen_vl.py
import torch import torch.nn as nn import torch.nn.functional as F from typing import Optional, Tuple import math from comfy.ldm.modules.attention import optimized_attention_for_device def process_qwen2vl_images( images: torch.Tensor, min_pixels: int = 3136, max_pixels: int = 12845056, patch_size: int = 14, temporal_patch_size: int = 2, merge_size: int = 2, image_mean: list = None, image_std: list = None, ): if image_mean is None: image_mean = [0.48145466, 0.4578275, 0.40821073] if image_std is None: image_std = [0.26862954, 0.26130258, 0.27577711] batch_size, height, width, channels = images.shape device = images.device # dtype = images.dtype images = images.permute(0, 3, 1, 2) grid_thw_list = [] img = images[0] factor = patch_size * merge_size h_bar = round(height / factor) * factor w_bar = round(width / factor) * factor if h_bar * w_bar > max_pixels: beta = math.sqrt((height * width) / max_pixels) h_bar = max(factor, math.floor(height / beta / factor) * factor) w_bar = max(factor, math.floor(width / beta / factor) * factor) elif h_bar * w_bar < min_pixels: beta = math.sqrt(min_pixels / (height * width)) h_bar = math.ceil(height * beta / factor) * factor w_bar = math.ceil(width * beta / factor) * factor img_resized = F.interpolate( img.unsqueeze(0), size=(h_bar, w_bar), mode='bilinear', align_corners=False ).squeeze(0) normalized = img_resized.clone() for c in range(3): normalized[c] = (img_resized[c] - image_mean[c]) / image_std[c] grid_h = h_bar // patch_size grid_w = w_bar // patch_size grid_thw = torch.tensor([1, grid_h, grid_w], device=device, dtype=torch.long) pixel_values = normalized grid_thw_list.append(grid_thw) image_grid_thw = torch.stack(grid_thw_list) grid_t = 1 channel = pixel_values.shape[0] pixel_values = pixel_values.unsqueeze(0).repeat(2, 1, 1, 1) patches = pixel_values.reshape( grid_t, temporal_patch_size, channel, grid_h // merge_size, merge_size, patch_size, grid_w // merge_size, merge_size, patch_size, ) patches = patches.permute(0, 3, 6, 4, 7, 2, 1, 5, 8) flatten_patches = patches.reshape( grid_t * grid_h * grid_w, channel * temporal_patch_size * patch_size * patch_size ) return flatten_patches, image_grid_thw class VisionPatchEmbed(nn.Module): def __init__( self, patch_size: int = 14, temporal_patch_size: int = 2, in_channels: int = 3, embed_dim: int = 3584, device=None, dtype=None, ops=None, ): super().__init__() self.patch_size = patch_size self.temporal_patch_size = temporal_patch_size self.in_channels = in_channels self.embed_dim = embed_dim kernel_size = [temporal_patch_size, patch_size, patch_size] self.proj = ops.Conv3d( in_channels, embed_dim, kernel_size=kernel_size, stride=kernel_size, bias=False, device=device, dtype=dtype ) def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: hidden_states = hidden_states.view( -1, self.in_channels, self.temporal_patch_size, self.patch_size, self.patch_size ) hidden_states = self.proj(hidden_states) return hidden_states.view(-1, self.embed_dim) def rotate_half(x): x1 = x[..., : x.shape[-1] // 2] x2 = x[..., x.shape[-1] // 2 :] return torch.cat((-x2, x1), dim=-1) def apply_rotary_pos_emb_vision(q, k, cos, sin): cos, sin = cos.unsqueeze(-2).float(), sin.unsqueeze(-2).float() q_embed = (q * cos) + (rotate_half(q) * sin) k_embed = (k * cos) + (rotate_half(k) * sin) return q_embed, k_embed class VisionRotaryEmbedding(nn.Module): def __init__(self, dim: int, theta: float = 10000.0): super().__init__() self.dim = dim self.theta = theta def forward(self, seqlen: int, device) -> torch.Tensor: inv_freq = 1.0 / (self.theta ** (torch.arange(0, self.dim, 2, dtype=torch.float, device=device) / self.dim)) seq = torch.arange(seqlen, device=inv_freq.device, dtype=inv_freq.dtype) freqs = torch.outer(seq, inv_freq) return freqs class PatchMerger(nn.Module): def __init__(self, dim: int, context_dim: int, spatial_merge_size: int = 2, device=None, dtype=None, ops=None): super().__init__() self.hidden_size = context_dim * (spatial_merge_size ** 2) self.ln_q = ops.RMSNorm(context_dim, eps=1e-6, device=device, dtype=dtype) self.mlp = nn.Sequential( ops.Linear(self.hidden_size, self.hidden_size, device=device, dtype=dtype), nn.GELU(), ops.Linear(self.hidden_size, dim, device=device, dtype=dtype), ) def forward(self, x: torch.Tensor) -> torch.Tensor: x = self.ln_q(x).reshape(-1, self.hidden_size) x = self.mlp(x) return x class VisionAttention(nn.Module): def __init__(self, hidden_size: int, num_heads: int, device=None, dtype=None, ops=None): super().__init__() self.hidden_size = hidden_size self.num_heads = num_heads self.head_dim = hidden_size // num_heads self.scaling = self.head_dim ** -0.5 self.qkv = ops.Linear(hidden_size, hidden_size * 3, bias=True, device=device, dtype=dtype) self.proj = ops.Linear(hidden_size, hidden_size, bias=True, device=device, dtype=dtype) def forward( self, hidden_states: torch.Tensor, position_embeddings: Optional[Tuple[torch.Tensor, torch.Tensor]] = None, cu_seqlens=None, optimized_attention=None, ) -> torch.Tensor: if hidden_states.dim() == 2: seq_length, _ = hidden_states.shape batch_size = 1 hidden_states = hidden_states.unsqueeze(0) else: batch_size, seq_length, _ = hidden_states.shape qkv = self.qkv(hidden_states) qkv = qkv.reshape(batch_size, seq_length, 3, self.num_heads, self.head_dim) query_states, key_states, value_states = qkv.reshape(seq_length, 3, self.num_heads, -1).permute(1, 0, 2, 3).unbind(0) if position_embeddings is not None: cos, sin = position_embeddings query_states, key_states = apply_rotary_pos_emb_vision(query_states, key_states, cos, sin) query_states = query_states.transpose(0, 1).unsqueeze(0) key_states = key_states.transpose(0, 1).unsqueeze(0) value_states = value_states.transpose(0, 1).unsqueeze(0) lengths = cu_seqlens[1:] - cu_seqlens[:-1] splits = [ torch.split(tensor, lengths.tolist(), dim=2) for tensor in (query_states, key_states, value_states) ] attn_outputs = [ optimized_attention(q, k, v, self.num_heads, skip_reshape=True) for q, k, v in zip(*splits) ] attn_output = torch.cat(attn_outputs, dim=1) attn_output = attn_output.reshape(seq_length, -1) attn_output = self.proj(attn_output) return attn_output class VisionMLP(nn.Module): def __init__(self, hidden_size: int, intermediate_size: int, device=None, dtype=None, ops=None): super().__init__() self.gate_proj = ops.Linear(hidden_size, intermediate_size, bias=True, device=device, dtype=dtype) self.up_proj = ops.Linear(hidden_size, intermediate_size, bias=True, device=device, dtype=dtype) self.down_proj = ops.Linear(intermediate_size, hidden_size, bias=True, device=device, dtype=dtype) self.act_fn = nn.SiLU() def forward(self, hidden_state): return self.down_proj(self.act_fn(self.gate_proj(hidden_state)) * self.up_proj(hidden_state)) class VisionBlock(nn.Module): def __init__(self, hidden_size: int, intermediate_size: int, num_heads: int, device=None, dtype=None, ops=None): super().__init__() self.norm1 = ops.RMSNorm(hidden_size, eps=1e-6, device=device, dtype=dtype) self.norm2 = ops.RMSNorm(hidden_size, eps=1e-6, device=device, dtype=dtype) self.attn = VisionAttention(hidden_size, num_heads, device=device, dtype=dtype, ops=ops) self.mlp = VisionMLP(hidden_size, intermediate_size, device=device, dtype=dtype, ops=ops) def forward( self, hidden_states: torch.Tensor, position_embeddings: Optional[Tuple[torch.Tensor, torch.Tensor]] = None, cu_seqlens=None, optimized_attention=None, ) -> torch.Tensor: residual = hidden_states hidden_states = self.norm1(hidden_states) hidden_states = self.attn(hidden_states, position_embeddings, cu_seqlens, optimized_attention) hidden_states = residual + hidden_states residual = hidden_states hidden_states = self.norm2(hidden_states) hidden_states = self.mlp(hidden_states) hidden_states = residual + hidden_states return hidden_states class Qwen2VLVisionTransformer(nn.Module): def __init__( self, hidden_size: int = 3584, output_hidden_size: int = 3584, intermediate_size: int = 3420, num_heads: int = 16, num_layers: int = 32, patch_size: int = 14, temporal_patch_size: int = 2, spatial_merge_size: int = 2, window_size: int = 112, device=None, dtype=None, ops=None ): super().__init__() self.hidden_size = hidden_size self.patch_size = patch_size self.spatial_merge_size = spatial_merge_size self.window_size = window_size self.fullatt_block_indexes = [7, 15, 23, 31] self.patch_embed = VisionPatchEmbed( patch_size=patch_size, temporal_patch_size=temporal_patch_size, in_channels=3, embed_dim=hidden_size, device=device, dtype=dtype, ops=ops, ) head_dim = hidden_size // num_heads self.rotary_pos_emb = VisionRotaryEmbedding(head_dim // 2) self.blocks = nn.ModuleList([ VisionBlock(hidden_size, intermediate_size, num_heads, device, dtype, ops) for _ in range(num_layers) ]) self.merger = PatchMerger( dim=output_hidden_size, context_dim=hidden_size, spatial_merge_size=spatial_merge_size, device=device, dtype=dtype, ops=ops, ) def get_window_index(self, grid_thw): window_index = [] cu_window_seqlens = [0] window_index_id = 0 vit_merger_window_size = self.window_size // self.spatial_merge_size // self.patch_size for grid_t, grid_h, grid_w in grid_thw: llm_grid_h = grid_h // self.spatial_merge_size llm_grid_w = grid_w // self.spatial_merge_size index = torch.arange(grid_t * llm_grid_h * llm_grid_w).reshape(grid_t, llm_grid_h, llm_grid_w) pad_h = vit_merger_window_size - llm_grid_h % vit_merger_window_size pad_w = vit_merger_window_size - llm_grid_w % vit_merger_window_size num_windows_h = (llm_grid_h + pad_h) // vit_merger_window_size num_windows_w = (llm_grid_w + pad_w) // vit_merger_window_size index_padded = F.pad(index, (0, pad_w, 0, pad_h), "constant", -100) index_padded = index_padded.reshape( grid_t, num_windows_h, vit_merger_window_size, num_windows_w, vit_merger_window_size, ) index_padded = index_padded.permute(0, 1, 3, 2, 4).reshape( grid_t, num_windows_h * num_windows_w, vit_merger_window_size, vit_merger_window_size, ) seqlens = (index_padded != -100).sum([2, 3]).reshape(-1) index_padded = index_padded.reshape(-1) index_new = index_padded[index_padded != -100] window_index.append(index_new + window_index_id) cu_seqlens_tmp = seqlens.cumsum(0) * self.spatial_merge_size * self.spatial_merge_size + cu_window_seqlens[-1] cu_window_seqlens.extend(cu_seqlens_tmp.tolist()) window_index_id += (grid_t * llm_grid_h * llm_grid_w).item() window_index = torch.cat(window_index, dim=0) return window_index, cu_window_seqlens def get_position_embeddings(self, grid_thw, device): pos_ids = [] for t, h, w in grid_thw: hpos_ids = torch.arange(h, device=device).unsqueeze(1).expand(-1, w) hpos_ids = hpos_ids.reshape( h // self.spatial_merge_size, self.spatial_merge_size, w // self.spatial_merge_size, self.spatial_merge_size, ) hpos_ids = hpos_ids.permute(0, 2, 1, 3).flatten() wpos_ids = torch.arange(w, device=device).unsqueeze(0).expand(h, -1) wpos_ids = wpos_ids.reshape( h // self.spatial_merge_size, self.spatial_merge_size, w // self.spatial_merge_size, self.spatial_merge_size, ) wpos_ids = wpos_ids.permute(0, 2, 1, 3).flatten() pos_ids.append(torch.stack([hpos_ids, wpos_ids], dim=-1).repeat(t, 1)) pos_ids = torch.cat(pos_ids, dim=0) max_grid_size = grid_thw[:, 1:].max() rotary_pos_emb_full = self.rotary_pos_emb(max_grid_size, device) return rotary_pos_emb_full[pos_ids].flatten(1) def forward( self, pixel_values: torch.Tensor, image_grid_thw: Optional[torch.Tensor] = None, ) -> torch.Tensor: optimized_attention = optimized_attention_for_device(pixel_values.device, mask=False, small_input=True) hidden_states = self.patch_embed(pixel_values) window_index, cu_window_seqlens = self.get_window_index(image_grid_thw) cu_window_seqlens = torch.tensor(cu_window_seqlens, device=hidden_states.device) cu_window_seqlens = torch.unique_consecutive(cu_window_seqlens) position_embeddings = self.get_position_embeddings(image_grid_thw, hidden_states.device) seq_len, _ = hidden_states.size() spatial_merge_unit = self.spatial_merge_size * self.spatial_merge_size hidden_states = hidden_states.reshape(seq_len // spatial_merge_unit, spatial_merge_unit, -1) hidden_states = hidden_states[window_index, :, :] hidden_states = hidden_states.reshape(seq_len, -1) position_embeddings = position_embeddings.reshape(seq_len // spatial_merge_unit, spatial_merge_unit, -1) position_embeddings = position_embeddings[window_index, :, :] position_embeddings = position_embeddings.reshape(seq_len, -1) position_embeddings = torch.cat((position_embeddings, position_embeddings), dim=-1) position_embeddings = (position_embeddings.cos(), position_embeddings.sin()) cu_seqlens = torch.repeat_interleave(image_grid_thw[:, 1] * image_grid_thw[:, 2], image_grid_thw[:, 0]).cumsum( dim=0, dtype=torch.int32, ) cu_seqlens = F.pad(cu_seqlens, (1, 0), value=0) for i, block in enumerate(self.blocks): if i in self.fullatt_block_indexes: cu_seqlens_now = cu_seqlens else: cu_seqlens_now = cu_window_seqlens hidden_states = block(hidden_states, position_embeddings, cu_seqlens_now, optimized_attention=optimized_attention) hidden_states = self.merger(hidden_states) return hidden_states
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/text_encoders/qwen_vl.py", "license": "GNU General Public License v3.0", "lines": 351, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_extras/nodes_qwen.py
import node_helpers import comfy.utils import math from typing_extensions import override from comfy_api.latest import ComfyExtension, io import comfy.model_management import torch import nodes class TextEncodeQwenImageEdit(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="TextEncodeQwenImageEdit", category="advanced/conditioning", inputs=[ io.Clip.Input("clip"), io.String.Input("prompt", multiline=True, dynamic_prompts=True), io.Vae.Input("vae", optional=True), io.Image.Input("image", optional=True), ], outputs=[ io.Conditioning.Output(), ], ) @classmethod def execute(cls, clip, prompt, vae=None, image=None) -> io.NodeOutput: ref_latent = None if image is None: images = [] else: samples = image.movedim(-1, 1) total = int(1024 * 1024) scale_by = math.sqrt(total / (samples.shape[3] * samples.shape[2])) width = round(samples.shape[3] * scale_by) height = round(samples.shape[2] * scale_by) s = comfy.utils.common_upscale(samples, width, height, "area", "disabled") image = s.movedim(1, -1) images = [image[:, :, :, :3]] if vae is not None: ref_latent = vae.encode(image[:, :, :, :3]) tokens = clip.tokenize(prompt, images=images) conditioning = clip.encode_from_tokens_scheduled(tokens) if ref_latent is not None: conditioning = node_helpers.conditioning_set_values(conditioning, {"reference_latents": [ref_latent]}, append=True) return io.NodeOutput(conditioning) class TextEncodeQwenImageEditPlus(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="TextEncodeQwenImageEditPlus", category="advanced/conditioning", inputs=[ io.Clip.Input("clip"), io.String.Input("prompt", multiline=True, dynamic_prompts=True), io.Vae.Input("vae", optional=True), io.Image.Input("image1", optional=True), io.Image.Input("image2", optional=True), io.Image.Input("image3", optional=True), ], outputs=[ io.Conditioning.Output(), ], ) @classmethod def execute(cls, clip, prompt, vae=None, image1=None, image2=None, image3=None) -> io.NodeOutput: ref_latents = [] images = [image1, image2, image3] images_vl = [] llama_template = "<|im_start|>system\nDescribe the key features of the input image (color, shape, size, texture, objects, background), then explain how the user's text instruction should alter or modify the image. Generate a new image that meets the user's requirements while maintaining consistency with the original input where appropriate.<|im_end|>\n<|im_start|>user\n{}<|im_end|>\n<|im_start|>assistant\n" image_prompt = "" for i, image in enumerate(images): if image is not None: samples = image.movedim(-1, 1) total = int(384 * 384) scale_by = math.sqrt(total / (samples.shape[3] * samples.shape[2])) width = round(samples.shape[3] * scale_by) height = round(samples.shape[2] * scale_by) s = comfy.utils.common_upscale(samples, width, height, "area", "disabled") images_vl.append(s.movedim(1, -1)) if vae is not None: total = int(1024 * 1024) scale_by = math.sqrt(total / (samples.shape[3] * samples.shape[2])) width = round(samples.shape[3] * scale_by / 8.0) * 8 height = round(samples.shape[2] * scale_by / 8.0) * 8 s = comfy.utils.common_upscale(samples, width, height, "area", "disabled") ref_latents.append(vae.encode(s.movedim(1, -1)[:, :, :, :3])) image_prompt += "Picture {}: <|vision_start|><|image_pad|><|vision_end|>".format(i + 1) tokens = clip.tokenize(image_prompt + prompt, images=images_vl, llama_template=llama_template) conditioning = clip.encode_from_tokens_scheduled(tokens) if len(ref_latents) > 0: conditioning = node_helpers.conditioning_set_values(conditioning, {"reference_latents": ref_latents}, append=True) return io.NodeOutput(conditioning) class EmptyQwenImageLayeredLatentImage(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="EmptyQwenImageLayeredLatentImage", display_name="Empty Qwen Image Layered Latent", category="latent/qwen", inputs=[ io.Int.Input("width", default=640, min=16, max=nodes.MAX_RESOLUTION, step=16), io.Int.Input("height", default=640, min=16, max=nodes.MAX_RESOLUTION, step=16), io.Int.Input("layers", default=3, min=0, max=nodes.MAX_RESOLUTION, step=1, advanced=True), io.Int.Input("batch_size", default=1, min=1, max=4096), ], outputs=[ io.Latent.Output(), ], ) @classmethod def execute(cls, width, height, layers, batch_size=1) -> io.NodeOutput: latent = torch.zeros([batch_size, 16, layers + 1, height // 8, width // 8], device=comfy.model_management.intermediate_device()) return io.NodeOutput({"samples": latent}) class QwenExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[io.ComfyNode]]: return [ TextEncodeQwenImageEdit, TextEncodeQwenImageEditPlus, EmptyQwenImageLayeredLatentImage, ] async def comfy_entrypoint() -> QwenExtension: return QwenExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_extras/nodes_qwen.py", "license": "GNU General Public License v3.0", "lines": 123, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy/context_windows.py
from __future__ import annotations from typing import TYPE_CHECKING, Callable import torch import numpy as np import collections from dataclasses import dataclass from abc import ABC, abstractmethod import logging import comfy.model_management import comfy.patcher_extension if TYPE_CHECKING: from comfy.model_base import BaseModel from comfy.model_patcher import ModelPatcher from comfy.controlnet import ControlBase class ContextWindowABC(ABC): def __init__(self): ... @abstractmethod def get_tensor(self, full: torch.Tensor) -> torch.Tensor: """ Get torch.Tensor applicable to current window. """ raise NotImplementedError("Not implemented.") @abstractmethod def add_window(self, full: torch.Tensor, to_add: torch.Tensor) -> torch.Tensor: """ Apply torch.Tensor of window to the full tensor, in place. Returns reference to updated full tensor, not a copy. """ raise NotImplementedError("Not implemented.") class ContextHandlerABC(ABC): def __init__(self): ... @abstractmethod def should_use_context(self, model: BaseModel, conds: list[list[dict]], x_in: torch.Tensor, timestep: torch.Tensor, model_options: dict[str]) -> bool: raise NotImplementedError("Not implemented.") @abstractmethod def get_resized_cond(self, cond_in: list[dict], x_in: torch.Tensor, window: ContextWindowABC, device=None) -> list: raise NotImplementedError("Not implemented.") @abstractmethod def execute(self, calc_cond_batch: Callable, model: BaseModel, conds: list[list[dict]], x_in: torch.Tensor, timestep: torch.Tensor, model_options: dict[str]): raise NotImplementedError("Not implemented.") class IndexListContextWindow(ContextWindowABC): def __init__(self, index_list: list[int], dim: int=0, total_frames: int=0): self.index_list = index_list self.context_length = len(index_list) self.dim = dim self.total_frames = total_frames self.center_ratio = (min(index_list) + max(index_list)) / (2 * total_frames) def get_tensor(self, full: torch.Tensor, device=None, dim=None, retain_index_list=[]) -> torch.Tensor: if dim is None: dim = self.dim if dim == 0 and full.shape[dim] == 1: return full idx = tuple([slice(None)] * dim + [self.index_list]) window = full[idx] if retain_index_list: idx = tuple([slice(None)] * dim + [retain_index_list]) window[idx] = full[idx] return window.to(device) def add_window(self, full: torch.Tensor, to_add: torch.Tensor, dim=None) -> torch.Tensor: if dim is None: dim = self.dim idx = tuple([slice(None)] * dim + [self.index_list]) full[idx] += to_add return full def get_region_index(self, num_regions: int) -> int: region_idx = int(self.center_ratio * num_regions) return min(max(region_idx, 0), num_regions - 1) class IndexListCallbacks: EVALUATE_CONTEXT_WINDOWS = "evaluate_context_windows" COMBINE_CONTEXT_WINDOW_RESULTS = "combine_context_window_results" EXECUTE_START = "execute_start" EXECUTE_CLEANUP = "execute_cleanup" RESIZE_COND_ITEM = "resize_cond_item" def init_callbacks(self): return {} @dataclass class ContextSchedule: name: str func: Callable @dataclass class ContextFuseMethod: name: str func: Callable ContextResults = collections.namedtuple("ContextResults", ['window_idx', 'sub_conds_out', 'sub_conds', 'window']) class IndexListContextHandler(ContextHandlerABC): def __init__(self, context_schedule: ContextSchedule, fuse_method: ContextFuseMethod, context_length: int=1, context_overlap: int=0, context_stride: int=1, closed_loop: bool=False, dim:int=0, freenoise: bool=False, cond_retain_index_list: list[int]=[], split_conds_to_windows: bool=False): self.context_schedule = context_schedule self.fuse_method = fuse_method self.context_length = context_length self.context_overlap = context_overlap self.context_stride = context_stride self.closed_loop = closed_loop self.dim = dim self._step = 0 self.freenoise = freenoise self.cond_retain_index_list = [int(x.strip()) for x in cond_retain_index_list.split(",")] if cond_retain_index_list else [] self.split_conds_to_windows = split_conds_to_windows self.callbacks = {} def should_use_context(self, model: BaseModel, conds: list[list[dict]], x_in: torch.Tensor, timestep: torch.Tensor, model_options: dict[str]) -> bool: # for now, assume first dim is batch - should have stored on BaseModel in actual implementation if x_in.size(self.dim) > self.context_length: logging.info(f"Using context windows {self.context_length} with overlap {self.context_overlap} for {x_in.size(self.dim)} frames.") if self.cond_retain_index_list: logging.info(f"Retaining original cond for indexes: {self.cond_retain_index_list}") return True return False def prepare_control_objects(self, control: ControlBase, device=None) -> ControlBase: if control.previous_controlnet is not None: self.prepare_control_objects(control.previous_controlnet, device) return control def get_resized_cond(self, cond_in: list[dict], x_in: torch.Tensor, window: IndexListContextWindow, device=None) -> list: if cond_in is None: return None # reuse or resize cond items to match context requirements resized_cond = [] # if multiple conds, split based on primary region if self.split_conds_to_windows and len(cond_in) > 1: region = window.get_region_index(len(cond_in)) logging.info(f"Splitting conds to windows; using region {region} for window {window.index_list[0]}-{window.index_list[-1]} with center ratio {window.center_ratio:.3f}") cond_in = [cond_in[region]] # cond object is a list containing a dict - outer list is irrelevant, so just loop through it for actual_cond in cond_in: resized_actual_cond = actual_cond.copy() # now we are in the inner dict - "pooled_output" is a tensor, "control" is a ControlBase object, "model_conds" is dictionary for key in actual_cond: try: cond_item = actual_cond[key] if isinstance(cond_item, torch.Tensor): # check that tensor is the expected length - x.size(0) if self.dim < cond_item.ndim and cond_item.size(self.dim) == x_in.size(self.dim): # if so, it's subsetting time - tell controls the expected indeces so they can handle them actual_cond_item = window.get_tensor(cond_item) resized_actual_cond[key] = actual_cond_item.to(device) else: resized_actual_cond[key] = cond_item.to(device) # look for control elif key == "control": resized_actual_cond[key] = self.prepare_control_objects(cond_item, device) elif isinstance(cond_item, dict): new_cond_item = cond_item.copy() # when in dictionary, look for tensors and CONDCrossAttn [comfy/conds.py] (has cond attr that is a tensor) for cond_key, cond_value in new_cond_item.items(): # Allow callbacks to handle custom conditioning items handled = False for callback in comfy.patcher_extension.get_all_callbacks( IndexListCallbacks.RESIZE_COND_ITEM, self.callbacks ): result = callback(cond_key, cond_value, window, x_in, device, new_cond_item) if result is not None: new_cond_item[cond_key] = result handled = True break if handled: continue if isinstance(cond_value, torch.Tensor): if (self.dim < cond_value.ndim and cond_value(self.dim) == x_in.size(self.dim)) or \ (cond_value.ndim < self.dim and cond_value.size(0) == x_in.size(self.dim)): new_cond_item[cond_key] = window.get_tensor(cond_value, device) # Handle audio_embed (temporal dim is 1) elif cond_key == "audio_embed" and hasattr(cond_value, "cond") and isinstance(cond_value.cond, torch.Tensor): audio_cond = cond_value.cond if audio_cond.ndim > 1 and audio_cond.size(1) == x_in.size(self.dim): new_cond_item[cond_key] = cond_value._copy_with(window.get_tensor(audio_cond, device, dim=1)) # Handle vace_context (temporal dim is 3) elif cond_key == "vace_context" and hasattr(cond_value, "cond") and isinstance(cond_value.cond, torch.Tensor): vace_cond = cond_value.cond if vace_cond.ndim >= 4 and vace_cond.size(3) == x_in.size(self.dim): sliced_vace = window.get_tensor(vace_cond, device, dim=3, retain_index_list=self.cond_retain_index_list) new_cond_item[cond_key] = cond_value._copy_with(sliced_vace) # if has cond that is a Tensor, check if needs to be subset elif hasattr(cond_value, "cond") and isinstance(cond_value.cond, torch.Tensor): if (self.dim < cond_value.cond.ndim and cond_value.cond.size(self.dim) == x_in.size(self.dim)) or \ (cond_value.cond.ndim < self.dim and cond_value.cond.size(0) == x_in.size(self.dim)): new_cond_item[cond_key] = cond_value._copy_with(window.get_tensor(cond_value.cond, device, retain_index_list=self.cond_retain_index_list)) elif cond_key == "num_video_frames": # for SVD new_cond_item[cond_key] = cond_value._copy_with(cond_value.cond) new_cond_item[cond_key].cond = window.context_length resized_actual_cond[key] = new_cond_item else: resized_actual_cond[key] = cond_item finally: del cond_item # just in case to prevent VRAM issues resized_cond.append(resized_actual_cond) return resized_cond def set_step(self, timestep: torch.Tensor, model_options: dict[str]): mask = torch.isclose(model_options["transformer_options"]["sample_sigmas"], timestep[0], rtol=0.0001) matches = torch.nonzero(mask) if torch.numel(matches) == 0: return # substep from multi-step sampler: keep self._step from the last full step self._step = int(matches[0].item()) def get_context_windows(self, model: BaseModel, x_in: torch.Tensor, model_options: dict[str]) -> list[IndexListContextWindow]: full_length = x_in.size(self.dim) # TODO: choose dim based on model context_windows = self.context_schedule.func(full_length, self, model_options) context_windows = [IndexListContextWindow(window, dim=self.dim, total_frames=full_length) for window in context_windows] return context_windows def execute(self, calc_cond_batch: Callable, model: BaseModel, conds: list[list[dict]], x_in: torch.Tensor, timestep: torch.Tensor, model_options: dict[str]): self.set_step(timestep, model_options) context_windows = self.get_context_windows(model, x_in, model_options) enumerated_context_windows = list(enumerate(context_windows)) conds_final = [torch.zeros_like(x_in) for _ in conds] if self.fuse_method.name == ContextFuseMethods.RELATIVE: counts_final = [torch.ones(get_shape_for_dim(x_in, self.dim), device=x_in.device) for _ in conds] else: counts_final = [torch.zeros(get_shape_for_dim(x_in, self.dim), device=x_in.device) for _ in conds] biases_final = [([0.0] * x_in.shape[self.dim]) for _ in conds] for callback in comfy.patcher_extension.get_all_callbacks(IndexListCallbacks.EXECUTE_START, self.callbacks): callback(self, model, x_in, conds, timestep, model_options) for enum_window in enumerated_context_windows: results = self.evaluate_context_windows(calc_cond_batch, model, x_in, conds, timestep, [enum_window], model_options) for result in results: self.combine_context_window_results(x_in, result.sub_conds_out, result.sub_conds, result.window, result.window_idx, len(enumerated_context_windows), timestep, conds_final, counts_final, biases_final) try: # finalize conds if self.fuse_method.name == ContextFuseMethods.RELATIVE: # relative is already normalized, so return as is del counts_final return conds_final else: # normalize conds via division by context usage counts for i in range(len(conds_final)): conds_final[i] /= counts_final[i] del counts_final return conds_final finally: for callback in comfy.patcher_extension.get_all_callbacks(IndexListCallbacks.EXECUTE_CLEANUP, self.callbacks): callback(self, model, x_in, conds, timestep, model_options) def evaluate_context_windows(self, calc_cond_batch: Callable, model: BaseModel, x_in: torch.Tensor, conds, timestep: torch.Tensor, enumerated_context_windows: list[tuple[int, IndexListContextWindow]], model_options, device=None, first_device=None): results: list[ContextResults] = [] for window_idx, window in enumerated_context_windows: # allow processing to end between context window executions for faster Cancel comfy.model_management.throw_exception_if_processing_interrupted() for callback in comfy.patcher_extension.get_all_callbacks(IndexListCallbacks.EVALUATE_CONTEXT_WINDOWS, self.callbacks): callback(self, model, x_in, conds, timestep, model_options, window_idx, window, model_options, device, first_device) # update exposed params model_options["transformer_options"]["context_window"] = window # get subsections of x, timestep, conds sub_x = window.get_tensor(x_in, device) sub_timestep = window.get_tensor(timestep, device, dim=0) sub_conds = [self.get_resized_cond(cond, x_in, window, device) for cond in conds] sub_conds_out = calc_cond_batch(model, sub_conds, sub_x, sub_timestep, model_options) if device is not None: for i in range(len(sub_conds_out)): sub_conds_out[i] = sub_conds_out[i].to(x_in.device) results.append(ContextResults(window_idx, sub_conds_out, sub_conds, window)) return results def combine_context_window_results(self, x_in: torch.Tensor, sub_conds_out, sub_conds, window: IndexListContextWindow, window_idx: int, total_windows: int, timestep: torch.Tensor, conds_final: list[torch.Tensor], counts_final: list[torch.Tensor], biases_final: list[torch.Tensor]): if self.fuse_method.name == ContextFuseMethods.RELATIVE: for pos, idx in enumerate(window.index_list): # bias is the influence of a specific index in relation to the whole context window bias = 1 - abs(idx - (window.index_list[0] + window.index_list[-1]) / 2) / ((window.index_list[-1] - window.index_list[0] + 1e-2) / 2) bias = max(1e-2, bias) # take weighted average relative to total bias of current idx for i in range(len(sub_conds_out)): bias_total = biases_final[i][idx] prev_weight = (bias_total / (bias_total + bias)) new_weight = (bias / (bias_total + bias)) # account for dims of tensors idx_window = tuple([slice(None)] * self.dim + [idx]) pos_window = tuple([slice(None)] * self.dim + [pos]) # apply new values conds_final[i][idx_window] = conds_final[i][idx_window] * prev_weight + sub_conds_out[i][pos_window] * new_weight biases_final[i][idx] = bias_total + bias else: # add conds and counts based on weights of fuse method weights = get_context_weights(window.context_length, x_in.shape[self.dim], window.index_list, self, sigma=timestep) weights_tensor = match_weights_to_dim(weights, x_in, self.dim, device=x_in.device) for i in range(len(sub_conds_out)): window.add_window(conds_final[i], sub_conds_out[i] * weights_tensor) window.add_window(counts_final[i], weights_tensor) for callback in comfy.patcher_extension.get_all_callbacks(IndexListCallbacks.COMBINE_CONTEXT_WINDOW_RESULTS, self.callbacks): callback(self, x_in, sub_conds_out, sub_conds, window, window_idx, total_windows, timestep, conds_final, counts_final, biases_final) def _prepare_sampling_wrapper(executor, model, noise_shape: torch.Tensor, *args, **kwargs): # limit noise_shape length to context_length for more accurate vram use estimation model_options = kwargs.get("model_options", None) if model_options is None: raise Exception("model_options not found in prepare_sampling_wrapper; this should never happen, something went wrong.") handler: IndexListContextHandler = model_options.get("context_handler", None) if handler is not None: noise_shape = list(noise_shape) noise_shape[handler.dim] = min(noise_shape[handler.dim], handler.context_length) return executor(model, noise_shape, *args, **kwargs) def create_prepare_sampling_wrapper(model: ModelPatcher): model.add_wrapper_with_key( comfy.patcher_extension.WrappersMP.PREPARE_SAMPLING, "ContextWindows_prepare_sampling", _prepare_sampling_wrapper ) def _sampler_sample_wrapper(executor, guider, sigmas, extra_args, callback, noise, *args, **kwargs): model_options = extra_args.get("model_options", None) if model_options is None: raise Exception("model_options not found in sampler_sample_wrapper; this should never happen, something went wrong.") handler: IndexListContextHandler = model_options.get("context_handler", None) if handler is None: raise Exception("context_handler not found in sampler_sample_wrapper; this should never happen, something went wrong.") if not handler.freenoise: return executor(guider, sigmas, extra_args, callback, noise, *args, **kwargs) noise = apply_freenoise(noise, handler.dim, handler.context_length, handler.context_overlap, extra_args["seed"]) return executor(guider, sigmas, extra_args, callback, noise, *args, **kwargs) def create_sampler_sample_wrapper(model: ModelPatcher): model.add_wrapper_with_key( comfy.patcher_extension.WrappersMP.SAMPLER_SAMPLE, "ContextWindows_sampler_sample", _sampler_sample_wrapper ) def match_weights_to_dim(weights: list[float], x_in: torch.Tensor, dim: int, device=None) -> torch.Tensor: total_dims = len(x_in.shape) weights_tensor = torch.Tensor(weights).to(device=device) for _ in range(dim): weights_tensor = weights_tensor.unsqueeze(0) for _ in range(total_dims - dim - 1): weights_tensor = weights_tensor.unsqueeze(-1) return weights_tensor def get_shape_for_dim(x_in: torch.Tensor, dim: int) -> list[int]: total_dims = len(x_in.shape) shape = [] for _ in range(dim): shape.append(1) shape.append(x_in.shape[dim]) for _ in range(total_dims - dim - 1): shape.append(1) return shape class ContextSchedules: UNIFORM_LOOPED = "looped_uniform" UNIFORM_STANDARD = "standard_uniform" STATIC_STANDARD = "standard_static" BATCHED = "batched" # from https://github.com/neggles/animatediff-cli/blob/main/src/animatediff/pipelines/context.py def create_windows_uniform_looped(num_frames: int, handler: IndexListContextHandler, model_options: dict[str]): windows = [] if num_frames < handler.context_length: windows.append(list(range(num_frames))) return windows context_stride = min(handler.context_stride, int(np.ceil(np.log2(num_frames / handler.context_length))) + 1) # obtain uniform windows as normal, looping and all for context_step in 1 << np.arange(context_stride): pad = int(round(num_frames * ordered_halving(handler._step))) for j in range( int(ordered_halving(handler._step) * context_step) + pad, num_frames + pad + (0 if handler.closed_loop else -handler.context_overlap), (handler.context_length * context_step - handler.context_overlap), ): windows.append([e % num_frames for e in range(j, j + handler.context_length * context_step, context_step)]) return windows def create_windows_uniform_standard(num_frames: int, handler: IndexListContextHandler, model_options: dict[str]): # unlike looped, uniform_straight does NOT allow windows that loop back to the beginning; # instead, they get shifted to the corresponding end of the frames. # in the case that a window (shifted or not) is identical to the previous one, it gets skipped. windows = [] if num_frames <= handler.context_length: windows.append(list(range(num_frames))) return windows context_stride = min(handler.context_stride, int(np.ceil(np.log2(num_frames / handler.context_length))) + 1) # first, obtain uniform windows as normal, looping and all for context_step in 1 << np.arange(context_stride): pad = int(round(num_frames * ordered_halving(handler._step))) for j in range( int(ordered_halving(handler._step) * context_step) + pad, num_frames + pad + (-handler.context_overlap), (handler.context_length * context_step - handler.context_overlap), ): windows.append([e % num_frames for e in range(j, j + handler.context_length * context_step, context_step)]) # now that windows are created, shift any windows that loop, and delete duplicate windows delete_idxs = [] win_i = 0 while win_i < len(windows): # if window is rolls over itself, need to shift it is_roll, roll_idx = does_window_roll_over(windows[win_i], num_frames) if is_roll: roll_val = windows[win_i][roll_idx] # roll_val might not be 0 for windows of higher strides shift_window_to_end(windows[win_i], num_frames=num_frames) # check if next window (cyclical) is missing roll_val if roll_val not in windows[(win_i+1) % len(windows)]: # need to insert new window here - just insert window starting at roll_val windows.insert(win_i+1, list(range(roll_val, roll_val + handler.context_length))) # delete window if it's not unique for pre_i in range(0, win_i): if windows[win_i] == windows[pre_i]: delete_idxs.append(win_i) break win_i += 1 # reverse delete_idxs so that they will be deleted in an order that doesn't break idx correlation delete_idxs.reverse() for i in delete_idxs: windows.pop(i) return windows def create_windows_static_standard(num_frames: int, handler: IndexListContextHandler, model_options: dict[str]): windows = [] if num_frames <= handler.context_length: windows.append(list(range(num_frames))) return windows # always return the same set of windows delta = handler.context_length - handler.context_overlap for start_idx in range(0, num_frames, delta): # if past the end of frames, move start_idx back to allow same context_length ending = start_idx + handler.context_length if ending >= num_frames: final_delta = ending - num_frames final_start_idx = start_idx - final_delta windows.append(list(range(final_start_idx, final_start_idx + handler.context_length))) break windows.append(list(range(start_idx, start_idx + handler.context_length))) return windows def create_windows_batched(num_frames: int, handler: IndexListContextHandler, model_options: dict[str]): windows = [] if num_frames <= handler.context_length: windows.append(list(range(num_frames))) return windows # always return the same set of windows; # no overlap, just cut up based on context_length; # last window size will be different if num_frames % opts.context_length != 0 for start_idx in range(0, num_frames, handler.context_length): windows.append(list(range(start_idx, min(start_idx + handler.context_length, num_frames)))) return windows def create_windows_default(num_frames: int, handler: IndexListContextHandler): return [list(range(num_frames))] CONTEXT_MAPPING = { ContextSchedules.UNIFORM_LOOPED: create_windows_uniform_looped, ContextSchedules.UNIFORM_STANDARD: create_windows_uniform_standard, ContextSchedules.STATIC_STANDARD: create_windows_static_standard, ContextSchedules.BATCHED: create_windows_batched, } def get_matching_context_schedule(context_schedule: str) -> ContextSchedule: func = CONTEXT_MAPPING.get(context_schedule, None) if func is None: raise ValueError(f"Unknown context_schedule '{context_schedule}'.") return ContextSchedule(context_schedule, func) def get_context_weights(length: int, full_length: int, idxs: list[int], handler: IndexListContextHandler, sigma: torch.Tensor=None): return handler.fuse_method.func(length, sigma=sigma, handler=handler, full_length=full_length, idxs=idxs) def create_weights_flat(length: int, **kwargs) -> list[float]: # weight is the same for all return [1.0] * length def create_weights_pyramid(length: int, **kwargs) -> list[float]: # weight is based on the distance away from the edge of the context window; # based on weighted average concept in FreeNoise paper if length % 2 == 0: max_weight = length // 2 weight_sequence = list(range(1, max_weight + 1, 1)) + list(range(max_weight, 0, -1)) else: max_weight = (length + 1) // 2 weight_sequence = list(range(1, max_weight, 1)) + [max_weight] + list(range(max_weight - 1, 0, -1)) return weight_sequence def create_weights_overlap_linear(length: int, full_length: int, idxs: list[int], handler: IndexListContextHandler, **kwargs): # based on code in Kijai's WanVideoWrapper: https://github.com/kijai/ComfyUI-WanVideoWrapper/blob/dbb2523b37e4ccdf45127e5ae33e31362f755c8e/nodes.py#L1302 # only expected overlap is given different weights weights_torch = torch.ones((length)) # blend left-side on all except first window if min(idxs) > 0: ramp_up = torch.linspace(1e-37, 1, handler.context_overlap) weights_torch[:handler.context_overlap] = ramp_up # blend right-side on all except last window if max(idxs) < full_length-1: ramp_down = torch.linspace(1, 1e-37, handler.context_overlap) weights_torch[-handler.context_overlap:] = ramp_down return weights_torch class ContextFuseMethods: FLAT = "flat" PYRAMID = "pyramid" RELATIVE = "relative" OVERLAP_LINEAR = "overlap-linear" LIST = [PYRAMID, FLAT, OVERLAP_LINEAR] LIST_STATIC = [PYRAMID, RELATIVE, FLAT, OVERLAP_LINEAR] FUSE_MAPPING = { ContextFuseMethods.FLAT: create_weights_flat, ContextFuseMethods.PYRAMID: create_weights_pyramid, ContextFuseMethods.RELATIVE: create_weights_pyramid, ContextFuseMethods.OVERLAP_LINEAR: create_weights_overlap_linear, } def get_matching_fuse_method(fuse_method: str) -> ContextFuseMethod: func = FUSE_MAPPING.get(fuse_method, None) if func is None: raise ValueError(f"Unknown fuse_method '{fuse_method}'.") return ContextFuseMethod(fuse_method, func) # Returns fraction that has denominator that is a power of 2 def ordered_halving(val): # get binary value, padded with 0s for 64 bits bin_str = f"{val:064b}" # flip binary value, padding included bin_flip = bin_str[::-1] # convert binary to int as_int = int(bin_flip, 2) # divide by 1 << 64, equivalent to 2**64, or 18446744073709551616, # or b10000000000000000000000000000000000000000000000000000000000000000 (1 with 64 zero's) return as_int / (1 << 64) def get_missing_indexes(windows: list[list[int]], num_frames: int) -> list[int]: all_indexes = list(range(num_frames)) for w in windows: for val in w: try: all_indexes.remove(val) except ValueError: pass return all_indexes def does_window_roll_over(window: list[int], num_frames: int) -> tuple[bool, int]: prev_val = -1 for i, val in enumerate(window): val = val % num_frames if val < prev_val: return True, i prev_val = val return False, -1 def shift_window_to_start(window: list[int], num_frames: int): start_val = window[0] for i in range(len(window)): # 1) subtract each element by start_val to move vals relative to the start of all frames # 2) add num_frames and take modulus to get adjusted vals window[i] = ((window[i] - start_val) + num_frames) % num_frames def shift_window_to_end(window: list[int], num_frames: int): # 1) shift window to start shift_window_to_start(window, num_frames) end_val = window[-1] end_delta = num_frames - end_val - 1 for i in range(len(window)): # 2) add end_delta to each val to slide windows to end window[i] = window[i] + end_delta # https://github.com/Kosinkadink/ComfyUI-AnimateDiff-Evolved/blob/90fb1331201a4b29488089e4fbffc0d82cc6d0a9/animatediff/sample_settings.py#L465 def apply_freenoise(noise: torch.Tensor, dim: int, context_length: int, context_overlap: int, seed: int): logging.info("Context windows: Applying FreeNoise") generator = torch.Generator(device='cpu').manual_seed(seed) latent_video_length = noise.shape[dim] delta = context_length - context_overlap for start_idx in range(0, latent_video_length - context_length, delta): place_idx = start_idx + context_length actual_delta = min(delta, latent_video_length - place_idx) if actual_delta <= 0: break list_idx = torch.randperm(actual_delta, generator=generator, device='cpu') + start_idx source_slice = [slice(None)] * noise.ndim source_slice[dim] = list_idx target_slice = [slice(None)] * noise.ndim target_slice[dim] = slice(place_idx, place_idx + actual_delta) noise[tuple(target_slice)] = noise[tuple(source_slice)] return noise
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/context_windows.py", "license": "GNU General Public License v3.0", "lines": 537, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_extras/nodes_context_windows.py
from __future__ import annotations from comfy_api.latest import ComfyExtension, io import comfy.context_windows import nodes class ContextWindowsManualNode(io.ComfyNode): @classmethod def define_schema(cls) -> io.Schema: return io.Schema( node_id="ContextWindowsManual", display_name="Context Windows (Manual)", category="context", description="Manually set context windows.", inputs=[ io.Model.Input("model", tooltip="The model to apply context windows to during sampling."), io.Int.Input("context_length", min=1, default=16, tooltip="The length of the context window.", advanced=True), io.Int.Input("context_overlap", min=0, default=4, tooltip="The overlap of the context window.", advanced=True), io.Combo.Input("context_schedule", options=[ comfy.context_windows.ContextSchedules.STATIC_STANDARD, comfy.context_windows.ContextSchedules.UNIFORM_STANDARD, comfy.context_windows.ContextSchedules.UNIFORM_LOOPED, comfy.context_windows.ContextSchedules.BATCHED, ], tooltip="The stride of the context window."), io.Int.Input("context_stride", min=1, default=1, tooltip="The stride of the context window; only applicable to uniform schedules.", advanced=True), io.Boolean.Input("closed_loop", default=False, tooltip="Whether to close the context window loop; only applicable to looped schedules."), io.Combo.Input("fuse_method", options=comfy.context_windows.ContextFuseMethods.LIST_STATIC, default=comfy.context_windows.ContextFuseMethods.PYRAMID, tooltip="The method to use to fuse the context windows."), io.Int.Input("dim", min=0, max=5, default=0, tooltip="The dimension to apply the context windows to."), io.Boolean.Input("freenoise", default=False, tooltip="Whether to apply FreeNoise noise shuffling, improves window blending."), #io.String.Input("cond_retain_index_list", default="", tooltip="List of latent indices to retain in the conditioning tensors for each window, for example setting this to '0' will use the initial start image for each window."), #io.Boolean.Input("split_conds_to_windows", default=False, tooltip="Whether to split multiple conditionings (created by ConditionCombine) to each window based on region index."), ], outputs=[ io.Model.Output(tooltip="The model with context windows applied during sampling."), ], is_experimental=True, ) @classmethod def execute(cls, model: io.Model.Type, context_length: int, context_overlap: int, context_schedule: str, context_stride: int, closed_loop: bool, fuse_method: str, dim: int, freenoise: bool, cond_retain_index_list: list[int]=[], split_conds_to_windows: bool=False) -> io.Model: model = model.clone() model.model_options["context_handler"] = comfy.context_windows.IndexListContextHandler( context_schedule=comfy.context_windows.get_matching_context_schedule(context_schedule), fuse_method=comfy.context_windows.get_matching_fuse_method(fuse_method), context_length=context_length, context_overlap=context_overlap, context_stride=context_stride, closed_loop=closed_loop, dim=dim, freenoise=freenoise, cond_retain_index_list=cond_retain_index_list, split_conds_to_windows=split_conds_to_windows ) # make memory usage calculation only take into account the context window latents comfy.context_windows.create_prepare_sampling_wrapper(model) if freenoise: # no other use for this wrapper at this time comfy.context_windows.create_sampler_sample_wrapper(model) return io.NodeOutput(model) class WanContextWindowsManualNode(ContextWindowsManualNode): @classmethod def define_schema(cls) -> io.Schema: schema = super().define_schema() schema.node_id = "WanContextWindowsManual" schema.display_name = "WAN Context Windows (Manual)" schema.description = "Manually set context windows for WAN-like models (dim=2)." schema.inputs = [ io.Model.Input("model", tooltip="The model to apply context windows to during sampling."), io.Int.Input("context_length", min=1, max=nodes.MAX_RESOLUTION, step=4, default=81, tooltip="The length of the context window.", advanced=True), io.Int.Input("context_overlap", min=0, default=30, tooltip="The overlap of the context window.", advanced=True), io.Combo.Input("context_schedule", options=[ comfy.context_windows.ContextSchedules.STATIC_STANDARD, comfy.context_windows.ContextSchedules.UNIFORM_STANDARD, comfy.context_windows.ContextSchedules.UNIFORM_LOOPED, comfy.context_windows.ContextSchedules.BATCHED, ], tooltip="The stride of the context window."), io.Int.Input("context_stride", min=1, default=1, tooltip="The stride of the context window; only applicable to uniform schedules.", advanced=True), io.Boolean.Input("closed_loop", default=False, tooltip="Whether to close the context window loop; only applicable to looped schedules."), io.Combo.Input("fuse_method", options=comfy.context_windows.ContextFuseMethods.LIST_STATIC, default=comfy.context_windows.ContextFuseMethods.PYRAMID, tooltip="The method to use to fuse the context windows."), io.Boolean.Input("freenoise", default=False, tooltip="Whether to apply FreeNoise noise shuffling, improves window blending."), #io.String.Input("cond_retain_index_list", default="", tooltip="List of latent indices to retain in the conditioning tensors for each window, for example setting this to '0' will use the initial start image for each window."), #io.Boolean.Input("split_conds_to_windows", default=False, tooltip="Whether to split multiple conditionings (created by ConditionCombine) to each window based on region index."), ] return schema @classmethod def execute(cls, model: io.Model.Type, context_length: int, context_overlap: int, context_schedule: str, context_stride: int, closed_loop: bool, fuse_method: str, freenoise: bool, cond_retain_index_list: list[int]=[], split_conds_to_windows: bool=False) -> io.Model: context_length = max(((context_length - 1) // 4) + 1, 1) # at least length 1 context_overlap = max(((context_overlap - 1) // 4) + 1, 0) # at least overlap 0 return super().execute(model, context_length, context_overlap, context_schedule, context_stride, closed_loop, fuse_method, dim=2, freenoise=freenoise, cond_retain_index_list=cond_retain_index_list, split_conds_to_windows=split_conds_to_windows) class ContextWindowsExtension(ComfyExtension): async def get_node_list(self) -> list[type[io.ComfyNode]]: return [ ContextWindowsManualNode, WanContextWindowsManualNode, ] def comfy_entrypoint(): return ContextWindowsExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_extras/nodes_context_windows.py", "license": "GNU General Public License v3.0", "lines": 95, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy/ldm/qwen_image/model.py
# https://github.com/QwenLM/Qwen-Image (Apache 2.0) import torch import torch.nn as nn import torch.nn.functional as F from typing import Optional, Tuple from einops import repeat from comfy.ldm.lightricks.model import TimestepEmbedding, Timesteps from comfy.ldm.modules.attention import optimized_attention_masked from comfy.ldm.flux.layers import EmbedND import comfy.ldm.common_dit import comfy.patcher_extension from comfy.ldm.flux.math import apply_rope1 class GELU(nn.Module): def __init__(self, dim_in: int, dim_out: int, approximate: str = "none", bias: bool = True, dtype=None, device=None, operations=None): super().__init__() self.proj = operations.Linear(dim_in, dim_out, bias=bias, dtype=dtype, device=device) self.approximate = approximate def forward(self, hidden_states): hidden_states = self.proj(hidden_states) hidden_states = F.gelu(hidden_states, approximate=self.approximate) return hidden_states class FeedForward(nn.Module): def __init__( self, dim: int, dim_out: Optional[int] = None, mult: int = 4, dropout: float = 0.0, inner_dim=None, bias: bool = True, dtype=None, device=None, operations=None ): super().__init__() if inner_dim is None: inner_dim = int(dim * mult) dim_out = dim_out if dim_out is not None else dim self.net = nn.ModuleList([]) self.net.append(GELU(dim, inner_dim, approximate="tanh", bias=bias, dtype=dtype, device=device, operations=operations)) self.net.append(nn.Dropout(dropout)) self.net.append(operations.Linear(inner_dim, dim_out, bias=bias, dtype=dtype, device=device)) def forward(self, hidden_states: torch.Tensor, *args, **kwargs) -> torch.Tensor: for module in self.net: hidden_states = module(hidden_states) return hidden_states def apply_rotary_emb(x, freqs_cis): if x.shape[1] == 0: return x t_ = x.reshape(*x.shape[:-1], -1, 1, 2) t_out = freqs_cis[..., 0] * t_[..., 0] + freqs_cis[..., 1] * t_[..., 1] return t_out.reshape(*x.shape) class QwenTimestepProjEmbeddings(nn.Module): def __init__(self, embedding_dim, pooled_projection_dim, use_additional_t_cond=False, dtype=None, device=None, operations=None): super().__init__() self.time_proj = Timesteps(num_channels=256, flip_sin_to_cos=True, downscale_freq_shift=0, scale=1000) self.timestep_embedder = TimestepEmbedding( in_channels=256, time_embed_dim=embedding_dim, dtype=dtype, device=device, operations=operations ) self.use_additional_t_cond = use_additional_t_cond if self.use_additional_t_cond: self.addition_t_embedding = operations.Embedding(2, embedding_dim, device=device, dtype=dtype) def forward(self, timestep, hidden_states, addition_t_cond=None): timesteps_proj = self.time_proj(timestep) timesteps_emb = self.timestep_embedder(timesteps_proj.to(dtype=hidden_states.dtype)) if self.use_additional_t_cond: if addition_t_cond is None: addition_t_cond = torch.zeros((timesteps_emb.shape[0]), device=timesteps_emb.device, dtype=torch.long) timesteps_emb += self.addition_t_embedding(addition_t_cond, out_dtype=timesteps_emb.dtype) return timesteps_emb class Attention(nn.Module): def __init__( self, query_dim: int, dim_head: int = 64, heads: int = 8, dropout: float = 0.0, bias: bool = False, eps: float = 1e-5, out_bias: bool = True, out_dim: int = None, out_context_dim: int = None, dtype=None, device=None, operations=None ): super().__init__() self.inner_dim = out_dim if out_dim is not None else dim_head * heads self.inner_kv_dim = self.inner_dim self.heads = heads self.dim_head = dim_head self.out_dim = out_dim if out_dim is not None else query_dim self.out_context_dim = out_context_dim if out_context_dim is not None else query_dim self.dropout = dropout # Q/K normalization self.norm_q = operations.RMSNorm(dim_head, eps=eps, elementwise_affine=True, dtype=dtype, device=device) self.norm_k = operations.RMSNorm(dim_head, eps=eps, elementwise_affine=True, dtype=dtype, device=device) self.norm_added_q = operations.RMSNorm(dim_head, eps=eps, dtype=dtype, device=device) self.norm_added_k = operations.RMSNorm(dim_head, eps=eps, dtype=dtype, device=device) # Image stream projections self.to_q = operations.Linear(query_dim, self.inner_dim, bias=bias, dtype=dtype, device=device) self.to_k = operations.Linear(query_dim, self.inner_kv_dim, bias=bias, dtype=dtype, device=device) self.to_v = operations.Linear(query_dim, self.inner_kv_dim, bias=bias, dtype=dtype, device=device) # Text stream projections self.add_q_proj = operations.Linear(query_dim, self.inner_dim, bias=bias, dtype=dtype, device=device) self.add_k_proj = operations.Linear(query_dim, self.inner_kv_dim, bias=bias, dtype=dtype, device=device) self.add_v_proj = operations.Linear(query_dim, self.inner_kv_dim, bias=bias, dtype=dtype, device=device) # Output projections self.to_out = nn.ModuleList([ operations.Linear(self.inner_dim, self.out_dim, bias=out_bias, dtype=dtype, device=device), nn.Dropout(dropout) ]) self.to_add_out = operations.Linear(self.inner_dim, self.out_context_dim, bias=out_bias, dtype=dtype, device=device) def forward( self, hidden_states: torch.FloatTensor, # Image stream encoder_hidden_states: torch.FloatTensor = None, # Text stream encoder_hidden_states_mask: torch.FloatTensor = None, attention_mask: Optional[torch.FloatTensor] = None, image_rotary_emb: Optional[torch.Tensor] = None, transformer_options={}, ) -> Tuple[torch.Tensor, torch.Tensor]: batch_size = hidden_states.shape[0] seq_img = hidden_states.shape[1] seq_txt = encoder_hidden_states.shape[1] # Project and reshape to BHND format (batch, heads, seq, dim) img_query = self.to_q(hidden_states).view(batch_size, seq_img, self.heads, -1).transpose(1, 2).contiguous() img_key = self.to_k(hidden_states).view(batch_size, seq_img, self.heads, -1).transpose(1, 2).contiguous() img_value = self.to_v(hidden_states).view(batch_size, seq_img, self.heads, -1).transpose(1, 2) txt_query = self.add_q_proj(encoder_hidden_states).view(batch_size, seq_txt, self.heads, -1).transpose(1, 2).contiguous() txt_key = self.add_k_proj(encoder_hidden_states).view(batch_size, seq_txt, self.heads, -1).transpose(1, 2).contiguous() txt_value = self.add_v_proj(encoder_hidden_states).view(batch_size, seq_txt, self.heads, -1).transpose(1, 2) img_query = self.norm_q(img_query) img_key = self.norm_k(img_key) txt_query = self.norm_added_q(txt_query) txt_key = self.norm_added_k(txt_key) joint_query = torch.cat([txt_query, img_query], dim=2) joint_key = torch.cat([txt_key, img_key], dim=2) joint_value = torch.cat([txt_value, img_value], dim=2) joint_query = apply_rope1(joint_query, image_rotary_emb) joint_key = apply_rope1(joint_key, image_rotary_emb) if encoder_hidden_states_mask is not None: attn_mask = torch.zeros((batch_size, 1, seq_txt + seq_img), dtype=hidden_states.dtype, device=hidden_states.device) attn_mask[:, 0, :seq_txt] = encoder_hidden_states_mask else: attn_mask = None joint_hidden_states = optimized_attention_masked(joint_query, joint_key, joint_value, self.heads, attn_mask, transformer_options=transformer_options, skip_reshape=True) txt_attn_output = joint_hidden_states[:, :seq_txt, :] img_attn_output = joint_hidden_states[:, seq_txt:, :] img_attn_output = self.to_out[0](img_attn_output) img_attn_output = self.to_out[1](img_attn_output) txt_attn_output = self.to_add_out(txt_attn_output) return img_attn_output, txt_attn_output class QwenImageTransformerBlock(nn.Module): def __init__( self, dim: int, num_attention_heads: int, attention_head_dim: int, eps: float = 1e-6, dtype=None, device=None, operations=None ): super().__init__() self.dim = dim self.num_attention_heads = num_attention_heads self.attention_head_dim = attention_head_dim self.img_mod = nn.Sequential( nn.SiLU(), operations.Linear(dim, 6 * dim, bias=True, dtype=dtype, device=device), ) self.img_norm1 = operations.LayerNorm(dim, elementwise_affine=False, eps=eps, dtype=dtype, device=device) self.img_norm2 = operations.LayerNorm(dim, elementwise_affine=False, eps=eps, dtype=dtype, device=device) self.img_mlp = FeedForward(dim=dim, dim_out=dim, dtype=dtype, device=device, operations=operations) self.txt_mod = nn.Sequential( nn.SiLU(), operations.Linear(dim, 6 * dim, bias=True, dtype=dtype, device=device), ) self.txt_norm1 = operations.LayerNorm(dim, elementwise_affine=False, eps=eps, dtype=dtype, device=device) self.txt_norm2 = operations.LayerNorm(dim, elementwise_affine=False, eps=eps, dtype=dtype, device=device) self.txt_mlp = FeedForward(dim=dim, dim_out=dim, dtype=dtype, device=device, operations=operations) self.attn = Attention( query_dim=dim, dim_head=attention_head_dim, heads=num_attention_heads, out_dim=dim, bias=True, eps=eps, dtype=dtype, device=device, operations=operations, ) def _apply_gate(self, x, y, gate, timestep_zero_index=None): if timestep_zero_index is not None: return y + torch.cat((x[:, :timestep_zero_index] * gate[0], x[:, timestep_zero_index:] * gate[1]), dim=1) else: return torch.addcmul(y, gate, x) def _modulate(self, x: torch.Tensor, mod_params: torch.Tensor, timestep_zero_index=None) -> Tuple[torch.Tensor, torch.Tensor]: shift, scale, gate = torch.chunk(mod_params, 3, dim=-1) if timestep_zero_index is not None: actual_batch = shift.size(0) // 2 shift, shift_0 = shift[:actual_batch], shift[actual_batch:] scale, scale_0 = scale[:actual_batch], scale[actual_batch:] gate, gate_0 = gate[:actual_batch], gate[actual_batch:] reg = torch.addcmul(shift.unsqueeze(1), x[:, :timestep_zero_index], 1 + scale.unsqueeze(1)) zero = torch.addcmul(shift_0.unsqueeze(1), x[:, timestep_zero_index:], 1 + scale_0.unsqueeze(1)) return torch.cat((reg, zero), dim=1), (gate.unsqueeze(1), gate_0.unsqueeze(1)) else: return torch.addcmul(shift.unsqueeze(1), x, 1 + scale.unsqueeze(1)), gate.unsqueeze(1) def forward( self, hidden_states: torch.Tensor, encoder_hidden_states: torch.Tensor, encoder_hidden_states_mask: torch.Tensor, temb: torch.Tensor, image_rotary_emb: Optional[Tuple[torch.Tensor, torch.Tensor]] = None, timestep_zero_index=None, transformer_options={}, ) -> Tuple[torch.Tensor, torch.Tensor]: img_mod_params = self.img_mod(temb) if timestep_zero_index is not None: temb = temb.chunk(2, dim=0)[0] txt_mod_params = self.txt_mod(temb) img_mod1, img_mod2 = img_mod_params.chunk(2, dim=-1) txt_mod1, txt_mod2 = txt_mod_params.chunk(2, dim=-1) img_modulated, img_gate1 = self._modulate(self.img_norm1(hidden_states), img_mod1, timestep_zero_index) del img_mod1 txt_modulated, txt_gate1 = self._modulate(self.txt_norm1(encoder_hidden_states), txt_mod1) del txt_mod1 img_attn_output, txt_attn_output = self.attn( hidden_states=img_modulated, encoder_hidden_states=txt_modulated, encoder_hidden_states_mask=encoder_hidden_states_mask, image_rotary_emb=image_rotary_emb, transformer_options=transformer_options, ) del img_modulated del txt_modulated hidden_states = self._apply_gate(img_attn_output, hidden_states, img_gate1, timestep_zero_index) encoder_hidden_states = encoder_hidden_states + txt_gate1 * txt_attn_output del img_attn_output del txt_attn_output del img_gate1 del txt_gate1 img_modulated2, img_gate2 = self._modulate(self.img_norm2(hidden_states), img_mod2, timestep_zero_index) hidden_states = self._apply_gate(self.img_mlp(img_modulated2), hidden_states, img_gate2, timestep_zero_index) txt_modulated2, txt_gate2 = self._modulate(self.txt_norm2(encoder_hidden_states), txt_mod2) encoder_hidden_states = torch.addcmul(encoder_hidden_states, txt_gate2, self.txt_mlp(txt_modulated2)) return encoder_hidden_states, hidden_states class LastLayer(nn.Module): def __init__( self, embedding_dim: int, conditioning_embedding_dim: int, elementwise_affine=False, eps=1e-6, bias=True, dtype=None, device=None, operations=None ): super().__init__() self.silu = nn.SiLU() self.linear = operations.Linear(conditioning_embedding_dim, embedding_dim * 2, bias=bias, dtype=dtype, device=device) self.norm = operations.LayerNorm(embedding_dim, eps, elementwise_affine=False, bias=bias, dtype=dtype, device=device) def forward(self, x: torch.Tensor, conditioning_embedding: torch.Tensor) -> torch.Tensor: emb = self.linear(self.silu(conditioning_embedding)) scale, shift = torch.chunk(emb, 2, dim=1) x = torch.addcmul(shift[:, None, :], self.norm(x), (1 + scale)[:, None, :]) return x class QwenImageTransformer2DModel(nn.Module): def __init__( self, patch_size: int = 2, in_channels: int = 64, out_channels: Optional[int] = 16, num_layers: int = 60, attention_head_dim: int = 128, num_attention_heads: int = 24, joint_attention_dim: int = 3584, pooled_projection_dim: int = 768, axes_dims_rope: Tuple[int, int, int] = (16, 56, 56), default_ref_method="index", image_model=None, final_layer=True, use_additional_t_cond=False, dtype=None, device=None, operations=None, ): super().__init__() self.dtype = dtype self.patch_size = patch_size self.in_channels = in_channels self.out_channels = out_channels or in_channels self.inner_dim = num_attention_heads * attention_head_dim self.default_ref_method = default_ref_method self.pe_embedder = EmbedND(dim=attention_head_dim, theta=10000, axes_dim=list(axes_dims_rope)) self.time_text_embed = QwenTimestepProjEmbeddings( embedding_dim=self.inner_dim, pooled_projection_dim=pooled_projection_dim, use_additional_t_cond=use_additional_t_cond, dtype=dtype, device=device, operations=operations ) self.txt_norm = operations.RMSNorm(joint_attention_dim, eps=1e-6, dtype=dtype, device=device) self.img_in = operations.Linear(in_channels, self.inner_dim, dtype=dtype, device=device) self.txt_in = operations.Linear(joint_attention_dim, self.inner_dim, dtype=dtype, device=device) self.transformer_blocks = nn.ModuleList([ QwenImageTransformerBlock( dim=self.inner_dim, num_attention_heads=num_attention_heads, attention_head_dim=attention_head_dim, dtype=dtype, device=device, operations=operations ) for _ in range(num_layers) ]) if self.default_ref_method == "index_timestep_zero": self.register_buffer("__index_timestep_zero__", torch.tensor([])) if final_layer: self.norm_out = LastLayer(self.inner_dim, self.inner_dim, dtype=dtype, device=device, operations=operations) self.proj_out = operations.Linear(self.inner_dim, patch_size * patch_size * self.out_channels, bias=True, dtype=dtype, device=device) def process_img(self, x, index=0, h_offset=0, w_offset=0): bs, c, t, h, w = x.shape patch_size = self.patch_size hidden_states = comfy.ldm.common_dit.pad_to_patch_size(x, (1, self.patch_size, self.patch_size)) orig_shape = hidden_states.shape hidden_states = hidden_states.view(orig_shape[0], orig_shape[1], orig_shape[-3], orig_shape[-2] // 2, 2, orig_shape[-1] // 2, 2) hidden_states = hidden_states.permute(0, 2, 3, 5, 1, 4, 6) hidden_states = hidden_states.reshape(orig_shape[0], orig_shape[-3] * (orig_shape[-2] // 2) * (orig_shape[-1] // 2), orig_shape[1] * 4) t_len = t h_len = ((h + (patch_size // 2)) // patch_size) w_len = ((w + (patch_size // 2)) // patch_size) h_offset = ((h_offset + (patch_size // 2)) // patch_size) w_offset = ((w_offset + (patch_size // 2)) // patch_size) img_ids = torch.zeros((t_len, h_len, w_len, 3), device=x.device) if t_len > 1: img_ids[:, :, :, 0] = img_ids[:, :, :, 0] + torch.linspace(0, t_len - 1, steps=t_len, device=x.device, dtype=x.dtype).unsqueeze(1).unsqueeze(1) else: img_ids[:, :, :, 0] = img_ids[:, :, :, 0] + index img_ids[:, :, :, 1] = img_ids[:, :, :, 1] + torch.linspace(h_offset, h_len - 1 + h_offset, steps=h_len, device=x.device, dtype=x.dtype).unsqueeze(1).unsqueeze(0) - (h_len // 2) img_ids[:, :, :, 2] = img_ids[:, :, :, 2] + torch.linspace(w_offset, w_len - 1 + w_offset, steps=w_len, device=x.device, dtype=x.dtype).unsqueeze(0).unsqueeze(0) - (w_len // 2) return hidden_states, repeat(img_ids, "t h w c -> b (t h w) c", b=bs), orig_shape def forward(self, x, timestep, context, attention_mask=None, ref_latents=None, additional_t_cond=None, transformer_options={}, **kwargs): return comfy.patcher_extension.WrapperExecutor.new_class_executor( self._forward, self, comfy.patcher_extension.get_all_wrappers(comfy.patcher_extension.WrappersMP.DIFFUSION_MODEL, transformer_options) ).execute(x, timestep, context, attention_mask, ref_latents, additional_t_cond, transformer_options, **kwargs) def _forward( self, x, timesteps, context, attention_mask=None, ref_latents=None, additional_t_cond=None, transformer_options={}, control=None, **kwargs ): timestep = timesteps encoder_hidden_states = context encoder_hidden_states_mask = attention_mask if encoder_hidden_states_mask is not None and not torch.is_floating_point(encoder_hidden_states_mask): encoder_hidden_states_mask = (encoder_hidden_states_mask - 1).to(x.dtype) * torch.finfo(x.dtype).max hidden_states, img_ids, orig_shape = self.process_img(x) num_embeds = hidden_states.shape[1] timestep_zero_index = None if ref_latents is not None: h = 0 w = 0 index = 0 ref_method = kwargs.get("ref_latents_method", self.default_ref_method) index_ref_method = (ref_method == "index") or (ref_method == "index_timestep_zero") negative_ref_method = ref_method == "negative_index" timestep_zero = ref_method == "index_timestep_zero" for ref in ref_latents: if index_ref_method: index += 1 h_offset = 0 w_offset = 0 elif negative_ref_method: index -= 1 h_offset = 0 w_offset = 0 else: index = 1 h_offset = 0 w_offset = 0 if ref.shape[-2] + h > ref.shape[-1] + w: w_offset = w else: h_offset = h h = max(h, ref.shape[-2] + h_offset) w = max(w, ref.shape[-1] + w_offset) kontext, kontext_ids, _ = self.process_img(ref, index=index, h_offset=h_offset, w_offset=w_offset) hidden_states = torch.cat([hidden_states, kontext], dim=1) img_ids = torch.cat([img_ids, kontext_ids], dim=1) if timestep_zero: if index > 0: timestep = torch.cat([timestep, timestep * 0], dim=0) timestep_zero_index = num_embeds txt_start = round(max(((x.shape[-1] + (self.patch_size // 2)) // self.patch_size) // 2, ((x.shape[-2] + (self.patch_size // 2)) // self.patch_size) // 2)) txt_ids = torch.arange(txt_start, txt_start + context.shape[1], device=x.device).reshape(1, -1, 1).repeat(x.shape[0], 1, 3) ids = torch.cat((txt_ids, img_ids), dim=1) image_rotary_emb = self.pe_embedder(ids).to(x.dtype).contiguous() del ids, txt_ids, img_ids hidden_states = self.img_in(hidden_states) encoder_hidden_states = self.txt_norm(encoder_hidden_states) encoder_hidden_states = self.txt_in(encoder_hidden_states) temb = self.time_text_embed(timestep, hidden_states, additional_t_cond) patches_replace = transformer_options.get("patches_replace", {}) patches = transformer_options.get("patches", {}) blocks_replace = patches_replace.get("dit", {}) transformer_options["total_blocks"] = len(self.transformer_blocks) transformer_options["block_type"] = "double" for i, block in enumerate(self.transformer_blocks): transformer_options["block_index"] = i if ("double_block", i) in blocks_replace: def block_wrap(args): out = {} out["txt"], out["img"] = block(hidden_states=args["img"], encoder_hidden_states=args["txt"], encoder_hidden_states_mask=encoder_hidden_states_mask, temb=args["vec"], image_rotary_emb=args["pe"], timestep_zero_index=timestep_zero_index, transformer_options=args["transformer_options"]) return out out = blocks_replace[("double_block", i)]({"img": hidden_states, "txt": encoder_hidden_states, "vec": temb, "pe": image_rotary_emb, "transformer_options": transformer_options}, {"original_block": block_wrap}) hidden_states = out["img"] encoder_hidden_states = out["txt"] else: encoder_hidden_states, hidden_states = block( hidden_states=hidden_states, encoder_hidden_states=encoder_hidden_states, encoder_hidden_states_mask=encoder_hidden_states_mask, temb=temb, image_rotary_emb=image_rotary_emb, timestep_zero_index=timestep_zero_index, transformer_options=transformer_options, ) if "double_block" in patches: for p in patches["double_block"]: out = p({"img": hidden_states, "txt": encoder_hidden_states, "x": x, "block_index": i, "transformer_options": transformer_options}) hidden_states = out["img"] encoder_hidden_states = out["txt"] if control is not None: # Controlnet control_i = control.get("input") if i < len(control_i): add = control_i[i] if add is not None: hidden_states[:, :add.shape[1]] += add if timestep_zero_index is not None: temb = temb.chunk(2, dim=0)[0] hidden_states = self.norm_out(hidden_states, temb) hidden_states = self.proj_out(hidden_states) hidden_states = hidden_states[:, :num_embeds].view(orig_shape[0], orig_shape[-3], orig_shape[-2] // 2, orig_shape[-1] // 2, orig_shape[1], 2, 2) hidden_states = hidden_states.permute(0, 4, 1, 2, 5, 3, 6) return hidden_states.reshape(orig_shape)[:, :, :, :x.shape[-2], :x.shape[-1]]
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/ldm/qwen_image/model.py", "license": "GNU General Public License v3.0", "lines": 461, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy/text_encoders/qwen_image.py
from transformers import Qwen2Tokenizer from comfy import sd1_clip import comfy.text_encoders.llama import os import torch import numbers class Qwen25_7BVLITokenizer(sd1_clip.SDTokenizer): def __init__(self, embedding_directory=None, tokenizer_data={}): tokenizer_path = os.path.join(os.path.dirname(os.path.realpath(__file__)), "qwen25_tokenizer") super().__init__(tokenizer_path, pad_with_end=False, embedding_size=3584, embedding_key='qwen25_7b', tokenizer_class=Qwen2Tokenizer, has_start_token=False, has_end_token=False, pad_to_max_length=False, max_length=99999999, min_length=1, pad_token=151643, tokenizer_data=tokenizer_data) class QwenImageTokenizer(sd1_clip.SD1Tokenizer): def __init__(self, embedding_directory=None, tokenizer_data={}): super().__init__(embedding_directory=embedding_directory, tokenizer_data=tokenizer_data, name="qwen25_7b", tokenizer=Qwen25_7BVLITokenizer) self.llama_template = "<|im_start|>system\nDescribe the image by detailing the color, shape, size, texture, quantity, text, spatial relationships of the objects and background:<|im_end|>\n<|im_start|>user\n{}<|im_end|>\n<|im_start|>assistant\n" self.llama_template_images = "<|im_start|>system\nDescribe the key features of the input image (color, shape, size, texture, objects, background), then explain how the user's text instruction should alter or modify the image. Generate a new image that meets the user's requirements while maintaining consistency with the original input where appropriate.<|im_end|>\n<|im_start|>user\n<|vision_start|><|image_pad|><|vision_end|>{}<|im_end|>\n<|im_start|>assistant\n" def tokenize_with_weights(self, text, return_word_ids=False, llama_template=None, images=[], prevent_empty_text=False, **kwargs): skip_template = False if text.startswith('<|im_start|>'): skip_template = True if text.startswith('<|start_header_id|>'): skip_template = True if prevent_empty_text and text == '': text = ' ' if skip_template: llama_text = text else: if llama_template is None: if len(images) > 0: llama_text = self.llama_template_images.format(text) else: llama_text = self.llama_template.format(text) else: llama_text = llama_template.format(text) tokens = super().tokenize_with_weights(llama_text, return_word_ids=return_word_ids, disable_weights=True, **kwargs) key_name = next(iter(tokens)) embed_count = 0 qwen_tokens = tokens[key_name] for r in qwen_tokens: for i in range(len(r)): if r[i][0] == 151655: if len(images) > embed_count: r[i] = ({"type": "image", "data": images[embed_count], "original_type": "image"},) + r[i][1:] embed_count += 1 return tokens class Qwen25_7BVLIModel(sd1_clip.SDClipModel): def __init__(self, device="cpu", layer="last", layer_idx=None, dtype=None, attention_mask=True, model_options={}): super().__init__(device=device, layer=layer, layer_idx=layer_idx, textmodel_json_config={}, dtype=dtype, special_tokens={"pad": 151643}, layer_norm_hidden_state=False, model_class=comfy.text_encoders.llama.Qwen25_7BVLI, enable_attention_masks=attention_mask, return_attention_masks=attention_mask, model_options=model_options) class QwenImageTEModel(sd1_clip.SD1ClipModel): def __init__(self, device="cpu", dtype=None, model_options={}): super().__init__(device=device, dtype=dtype, name="qwen25_7b", clip_model=Qwen25_7BVLIModel, model_options=model_options) def encode_token_weights(self, token_weight_pairs, template_end=-1): out, pooled, extra = super().encode_token_weights(token_weight_pairs) tok_pairs = token_weight_pairs["qwen25_7b"][0] count_im_start = 0 if template_end == -1: for i, v in enumerate(tok_pairs): elem = v[0] if not torch.is_tensor(elem): if isinstance(elem, numbers.Integral): if elem == 151644 and count_im_start < 2: template_end = i count_im_start += 1 if out.shape[1] > (template_end + 3): if tok_pairs[template_end + 1][0] == 872: if tok_pairs[template_end + 2][0] == 198: template_end += 3 out = out[:, template_end:] extra["attention_mask"] = extra["attention_mask"][:, template_end:] if extra["attention_mask"].sum() == torch.numel(extra["attention_mask"]): extra.pop("attention_mask") # attention mask is useless if no masked elements return out, pooled, extra def te(dtype_llama=None, llama_quantization_metadata=None): class QwenImageTEModel_(QwenImageTEModel): def __init__(self, device="cpu", dtype=None, model_options={}): if llama_quantization_metadata is not None: model_options = model_options.copy() model_options["quantization_metadata"] = llama_quantization_metadata if dtype_llama is not None: dtype = dtype_llama super().__init__(device=device, dtype=dtype, model_options=model_options) return QwenImageTEModel_
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/text_encoders/qwen_image.py", "license": "GNU General Public License v3.0", "lines": 81, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_api/latest/_io.py
from __future__ import annotations import copy import inspect from abc import ABC, abstractmethod from collections import Counter from collections.abc import Iterable from dataclasses import asdict, dataclass, field from enum import Enum from typing import Any, Callable, Literal, TypedDict, TypeVar, TYPE_CHECKING from typing_extensions import NotRequired, final # used for type hinting import torch if TYPE_CHECKING: from spandrel import ImageModelDescriptor from comfy.clip_vision import ClipVisionModel from comfy.clip_vision import Output as ClipVisionOutput_ from comfy.controlnet import ControlNet from comfy.hooks import HookGroup, HookKeyframeGroup from comfy.model_patcher import ModelPatcher from comfy.samplers import CFGGuider, Sampler from comfy.sd import CLIP, VAE from comfy.sd import StyleModel as StyleModel_ from comfy_api.input import VideoInput from comfy_api.internal import (_ComfyNodeInternal, _NodeOutputInternal, classproperty, copy_class, first_real_override, is_class, prune_dict, shallow_clone_class) from comfy_execution.graph_utils import ExecutionBlocker from ._util import MESH, VOXEL, SVG as _SVG, File3D class FolderType(str, Enum): input = "input" output = "output" temp = "temp" class UploadType(str, Enum): image = "image_upload" audio = "audio_upload" video = "video_upload" model = "file_upload" class RemoteOptions: def __init__(self, route: str, refresh_button: bool, control_after_refresh: Literal["first", "last"]="first", timeout: int=None, max_retries: int=None, refresh: int=None): self.route = route """The route to the remote source.""" self.refresh_button = refresh_button """Specifies whether to show a refresh button in the UI below the widget.""" self.control_after_refresh = control_after_refresh """Specifies the control after the refresh button is clicked. If "first", the first item will be automatically selected, and so on.""" self.timeout = timeout """The maximum amount of time to wait for a response from the remote source in milliseconds.""" self.max_retries = max_retries """The maximum number of retries before aborting the request.""" self.refresh = refresh """The TTL of the remote input's value in milliseconds. Specifies the interval at which the remote input's value is refreshed.""" def as_dict(self): return prune_dict({ "route": self.route, "refresh_button": self.refresh_button, "control_after_refresh": self.control_after_refresh, "timeout": self.timeout, "max_retries": self.max_retries, "refresh": self.refresh, }) class NumberDisplay(str, Enum): number = "number" slider = "slider" gradient_slider = "gradientslider" class ControlAfterGenerate(str, Enum): fixed = "fixed" increment = "increment" decrement = "decrement" randomize = "randomize" class _ComfyType(ABC): Type = Any io_type: str = None # NOTE: this is a workaround to make the decorator return the correct type T = TypeVar("T", bound=type) def comfytype(io_type: str, **kwargs): ''' Decorator to mark nested classes as ComfyType; io_type will be bound to the class. A ComfyType may have the following attributes: - Type = <type hint here> - class Input(Input): ... - class Output(Output): ... ''' def decorator(cls: T) -> T: if isinstance(cls, _ComfyType) or issubclass(cls, _ComfyType): # clone Input and Output classes to avoid modifying the original class new_cls = cls if hasattr(new_cls, "Input"): new_cls.Input = copy_class(new_cls.Input) if hasattr(new_cls, "Output"): new_cls.Output = copy_class(new_cls.Output) else: # copy class attributes except for special ones that shouldn't be in type() cls_dict = { k: v for k, v in cls.__dict__.items() if k not in ('__dict__', '__weakref__', '__module__', '__doc__') } # new class new_cls: ComfyTypeIO = type( cls.__name__, (cls, ComfyTypeIO), cls_dict ) # metadata preservation new_cls.__module__ = cls.__module__ new_cls.__doc__ = cls.__doc__ # assign ComfyType attributes, if needed new_cls.io_type = io_type if hasattr(new_cls, "Input") and new_cls.Input is not None: new_cls.Input.Parent = new_cls if hasattr(new_cls, "Output") and new_cls.Output is not None: new_cls.Output.Parent = new_cls return new_cls return decorator def Custom(io_type: str) -> type[ComfyTypeIO]: '''Create a ComfyType for a custom io_type.''' @comfytype(io_type=io_type) class CustomComfyType(ComfyTypeIO): ... return CustomComfyType class _IO_V3: ''' Base class for V3 Inputs and Outputs. ''' Parent: _ComfyType = None def __init__(self): pass def validate(self): pass @property def io_type(self): return self.Parent.io_type @property def Type(self): return self.Parent.Type class Input(_IO_V3): ''' Base class for a V3 Input. ''' def __init__(self, id: str, display_name: str=None, optional=False, tooltip: str=None, lazy: bool=None, extra_dict=None, raw_link: bool=None, advanced: bool=None): super().__init__() self.id = id self.display_name = display_name self.optional = optional self.tooltip = tooltip self.lazy = lazy self.extra_dict = extra_dict if extra_dict is not None else {} self.rawLink = raw_link self.advanced = advanced def as_dict(self): return prune_dict({ "display_name": self.display_name, "optional": self.optional, "tooltip": self.tooltip, "lazy": self.lazy, "rawLink": self.rawLink, "advanced": self.advanced, }) | prune_dict(self.extra_dict) def get_io_type(self): return self.io_type def get_all(self) -> list[Input]: return [self] class WidgetInput(Input): ''' Base class for a V3 Input with widget. ''' def __init__(self, id: str, display_name: str=None, optional=False, tooltip: str=None, lazy: bool=None, default: Any=None, socketless: bool=None, widget_type: str=None, force_input: bool=None, extra_dict=None, raw_link: bool=None, advanced: bool=None): super().__init__(id, display_name, optional, tooltip, lazy, extra_dict, raw_link, advanced) self.default = default self.socketless = socketless self.widget_type = widget_type self.force_input = force_input def as_dict(self): return super().as_dict() | prune_dict({ "default": self.default, "socketless": self.socketless, "widgetType": self.widget_type, "forceInput": self.force_input, }) def get_io_type(self): return self.widget_type if self.widget_type is not None else super().get_io_type() class Output(_IO_V3): def __init__(self, id: str=None, display_name: str=None, tooltip: str=None, is_output_list=False): self.id = id self.display_name = display_name if display_name else id self.tooltip = tooltip self.is_output_list = is_output_list def as_dict(self): display_name = self.display_name if self.display_name else self.id return prune_dict({ "display_name": display_name, "tooltip": self.tooltip, "is_output_list": self.is_output_list, }) def get_io_type(self): return self.io_type class ComfyTypeI(_ComfyType): '''ComfyType subclass that only has a default Input class - intended for types that only have Inputs.''' class Input(Input): ... class ComfyTypeIO(ComfyTypeI): '''ComfyType subclass that has default Input and Output classes; useful for types with both Inputs and Outputs.''' class Output(Output): ... @comfytype(io_type="BOOLEAN") class Boolean(ComfyTypeIO): Type = bool class Input(WidgetInput): '''Boolean input.''' def __init__(self, id: str, display_name: str=None, optional=False, tooltip: str=None, lazy: bool=None, default: bool=None, label_on: str=None, label_off: str=None, socketless: bool=None, force_input: bool=None, extra_dict=None, raw_link: bool=None, advanced: bool=None): super().__init__(id, display_name, optional, tooltip, lazy, default, socketless, None, force_input, extra_dict, raw_link, advanced) self.label_on = label_on self.label_off = label_off self.default: bool def as_dict(self): return super().as_dict() | prune_dict({ "label_on": self.label_on, "label_off": self.label_off, }) @comfytype(io_type="INT") class Int(ComfyTypeIO): Type = int class Input(WidgetInput): '''Integer input.''' def __init__(self, id: str, display_name: str=None, optional=False, tooltip: str=None, lazy: bool=None, default: int=None, min: int=None, max: int=None, step: int=None, control_after_generate: bool | ControlAfterGenerate=None, display_mode: NumberDisplay=None, socketless: bool=None, force_input: bool=None, extra_dict=None, raw_link: bool=None, advanced: bool=None): super().__init__(id, display_name, optional, tooltip, lazy, default, socketless, None, force_input, extra_dict, raw_link, advanced) self.min = min self.max = max self.step = step self.control_after_generate = control_after_generate self.display_mode = display_mode self.default: int def as_dict(self): return super().as_dict() | prune_dict({ "min": self.min, "max": self.max, "step": self.step, "control_after_generate": self.control_after_generate, "display": self.display_mode.value if self.display_mode else None, }) @comfytype(io_type="FLOAT") class Float(ComfyTypeIO): Type = float class Input(WidgetInput): '''Float input.''' def __init__(self, id: str, display_name: str=None, optional=False, tooltip: str=None, lazy: bool=None, default: float=None, min: float=None, max: float=None, step: float=None, round: float=None, display_mode: NumberDisplay=None, gradient_stops: list[list[float]]=None, socketless: bool=None, force_input: bool=None, extra_dict=None, raw_link: bool=None, advanced: bool=None): super().__init__(id, display_name, optional, tooltip, lazy, default, socketless, None, force_input, extra_dict, raw_link, advanced) self.min = min self.max = max self.step = step self.round = round self.display_mode = display_mode self.gradient_stops = gradient_stops self.default: float def as_dict(self): return super().as_dict() | prune_dict({ "min": self.min, "max": self.max, "step": self.step, "round": self.round, "display": self.display_mode, "gradient_stops": self.gradient_stops, }) @comfytype(io_type="STRING") class String(ComfyTypeIO): Type = str class Input(WidgetInput): '''String input.''' def __init__(self, id: str, display_name: str=None, optional=False, tooltip: str=None, lazy: bool=None, multiline=False, placeholder: str=None, default: str=None, dynamic_prompts: bool=None, socketless: bool=None, force_input: bool=None, extra_dict=None, raw_link: bool=None, advanced: bool=None): super().__init__(id, display_name, optional, tooltip, lazy, default, socketless, None, force_input, extra_dict, raw_link, advanced) self.multiline = multiline self.placeholder = placeholder self.dynamic_prompts = dynamic_prompts self.default: str def as_dict(self): return super().as_dict() | prune_dict({ "multiline": self.multiline, "placeholder": self.placeholder, "dynamicPrompts": self.dynamic_prompts, }) @comfytype(io_type="COMBO") class Combo(ComfyTypeIO): Type = str class Input(WidgetInput): """Combo input (dropdown).""" Type = str def __init__( self, id: str, options: list[str] | list[int] | type[Enum] = None, display_name: str=None, optional=False, tooltip: str=None, lazy: bool=None, default: str | int | Enum = None, control_after_generate: bool | ControlAfterGenerate=None, upload: UploadType=None, image_folder: FolderType=None, remote: RemoteOptions=None, socketless: bool=None, extra_dict=None, raw_link: bool=None, advanced: bool=None, ): if isinstance(options, type) and issubclass(options, Enum): options = [v.value for v in options] if isinstance(default, Enum): default = default.value super().__init__(id, display_name, optional, tooltip, lazy, default, socketless, None, None, extra_dict, raw_link, advanced) self.multiselect = False self.options = options self.control_after_generate = control_after_generate self.upload = upload self.image_folder = image_folder self.remote = remote self.default: str def as_dict(self): return super().as_dict() | prune_dict({ "multiselect": self.multiselect, "options": self.options, "control_after_generate": self.control_after_generate, **({self.upload.value: True} if self.upload is not None else {}), "image_folder": self.image_folder.value if self.image_folder else None, "remote": self.remote.as_dict() if self.remote else None, }) class Output(Output): def __init__(self, id: str=None, display_name: str=None, options: list[str]=None, tooltip: str=None, is_output_list=False): super().__init__(id, display_name, tooltip, is_output_list) self.options = options if options is not None else [] @comfytype(io_type="COMBO") class MultiCombo(ComfyTypeI): '''Multiselect Combo input (dropdown for selecting potentially more than one value).''' # TODO: something is wrong with the serialization, frontend does not recognize it as multiselect Type = list[str] class Input(Combo.Input): def __init__(self, id: str, options: list[str], display_name: str=None, optional=False, tooltip: str=None, lazy: bool=None, default: list[str]=None, placeholder: str=None, chip: bool=None, control_after_generate: bool | ControlAfterGenerate=None, socketless: bool=None, extra_dict=None, raw_link: bool=None, advanced: bool=None): super().__init__(id, options, display_name, optional, tooltip, lazy, default, control_after_generate, socketless=socketless, extra_dict=extra_dict, raw_link=raw_link, advanced=advanced) self.multiselect = True self.placeholder = placeholder self.chip = chip self.default: list[str] def as_dict(self): to_return = super().as_dict() | prune_dict({ "multi_select": self.multiselect, "placeholder": self.placeholder, "chip": self.chip, }) return to_return @comfytype(io_type="IMAGE") class Image(ComfyTypeIO): Type = torch.Tensor @comfytype(io_type="WAN_CAMERA_EMBEDDING") class WanCameraEmbedding(ComfyTypeIO): Type = torch.Tensor @comfytype(io_type="WEBCAM") class Webcam(ComfyTypeIO): Type = str class Input(WidgetInput): """Webcam input.""" Type = str def __init__( self, id: str, display_name: str=None, optional=False, tooltip: str=None, lazy: bool=None, default: str=None, socketless: bool=None, extra_dict=None, raw_link: bool=None, advanced: bool=None ): super().__init__(id, display_name, optional, tooltip, lazy, default, socketless, None, None, extra_dict, raw_link, advanced) @comfytype(io_type="MASK") class Mask(ComfyTypeIO): Type = torch.Tensor @comfytype(io_type="LATENT") class Latent(ComfyTypeIO): '''Latents are stored as a dictionary.''' class LatentDict(TypedDict): samples: torch.Tensor '''Latent tensors.''' noise_mask: NotRequired[torch.Tensor] batch_index: NotRequired[list[int]] type: NotRequired[str] '''Only needed if dealing with these types: audio, hunyuan3dv2''' Type = LatentDict @comfytype(io_type="CONDITIONING") class Conditioning(ComfyTypeIO): class PooledDict(TypedDict): pooled_output: torch.Tensor '''Pooled output from CLIP.''' control: NotRequired[ControlNet] '''ControlNet to apply to conditioning.''' control_apply_to_uncond: NotRequired[bool] '''Whether to apply ControlNet to matching negative conditioning at sample time, if applicable.''' cross_attn_controlnet: NotRequired[torch.Tensor] '''CrossAttn from CLIP to use for controlnet only.''' pooled_output_controlnet: NotRequired[torch.Tensor] '''Pooled output from CLIP to use for controlnet only.''' gligen: NotRequired[tuple[str, Gligen, list[tuple[torch.Tensor, int, ...]]]] '''GLIGEN to apply to conditioning.''' area: NotRequired[tuple[int, ...] | tuple[str, float, ...]] '''Set area of conditioning. First half of values apply to dimensions, the second half apply to coordinates. By default, the dimensions are based on total pixel amount, but the first value can be set to "percentage" to use a percentage of the image size instead. (1024, 1024, 0, 0) would apply conditioning to the top-left 1024x1024 pixels. ("percentage", 0.5, 0.5, 0, 0) would apply conditioning to the top-left 50% of the image.''' # TODO: verify its actually top-left strength: NotRequired[float] '''Strength of conditioning. Default strength is 1.0.''' mask: NotRequired[torch.Tensor] '''Mask to apply conditioning to.''' mask_strength: NotRequired[float] '''Strength of conditioning mask. Default strength is 1.0.''' set_area_to_bounds: NotRequired[bool] '''Whether conditioning mask should determine bounds of area - if set to false, latents are sampled at full resolution and result is applied in mask.''' concat_latent_image: NotRequired[torch.Tensor] '''Used for inpainting and specific models.''' concat_mask: NotRequired[torch.Tensor] '''Used for inpainting and specific models.''' concat_image: NotRequired[torch.Tensor] '''Used by SD_4XUpscale_Conditioning.''' noise_augmentation: NotRequired[float] '''Used by SD_4XUpscale_Conditioning.''' hooks: NotRequired[HookGroup] '''Applies hooks to conditioning.''' default: NotRequired[bool] '''Whether to this conditioning is 'default'; default conditioning gets applied to any areas of the image that have no masks/areas applied, assuming at least one area/mask is present during sampling.''' start_percent: NotRequired[float] '''Determines relative step to begin applying conditioning, expressed as a float between 0.0 and 1.0.''' end_percent: NotRequired[float] '''Determines relative step to end applying conditioning, expressed as a float between 0.0 and 1.0.''' clip_start_percent: NotRequired[float] '''Internal variable for conditioning scheduling - start of application, expressed as a float between 0.0 and 1.0.''' clip_end_percent: NotRequired[float] '''Internal variable for conditioning scheduling - end of application, expressed as a float between 0.0 and 1.0.''' attention_mask: NotRequired[torch.Tensor] '''Masks text conditioning; used by StyleModel among others.''' attention_mask_img_shape: NotRequired[tuple[int, ...]] '''Masks text conditioning; used by StyleModel among others.''' unclip_conditioning: NotRequired[list[dict]] '''Used by unCLIP.''' conditioning_lyrics: NotRequired[torch.Tensor] '''Used by AceT5Model.''' seconds_start: NotRequired[float] '''Used by StableAudio.''' seconds_total: NotRequired[float] '''Used by StableAudio.''' lyrics_strength: NotRequired[float] '''Used by AceStepAudio.''' width: NotRequired[int] '''Used by certain models (e.g. CLIPTextEncodeSDXL/Refiner, PixArtAlpha).''' height: NotRequired[int] '''Used by certain models (e.g. CLIPTextEncodeSDXL/Refiner, PixArtAlpha).''' aesthetic_score: NotRequired[float] '''Used by CLIPTextEncodeSDXL/Refiner.''' crop_w: NotRequired[int] '''Used by CLIPTextEncodeSDXL.''' crop_h: NotRequired[int] '''Used by CLIPTextEncodeSDXL.''' target_width: NotRequired[int] '''Used by CLIPTextEncodeSDXL.''' target_height: NotRequired[int] '''Used by CLIPTextEncodeSDXL.''' reference_latents: NotRequired[list[torch.Tensor]] '''Used by ReferenceLatent.''' guidance: NotRequired[float] '''Used by Flux-like models with guidance embed.''' guiding_frame_index: NotRequired[int] '''Used by Hunyuan ImageToVideo.''' ref_latent: NotRequired[torch.Tensor] '''Used by Hunyuan ImageToVideo.''' keyframe_idxs: NotRequired[list[int]] '''Used by LTXV.''' frame_rate: NotRequired[float] '''Used by LTXV.''' stable_cascade_prior: NotRequired[torch.Tensor] '''Used by StableCascade.''' elevation: NotRequired[list[float]] '''Used by SV3D.''' azimuth: NotRequired[list[float]] '''Used by SV3D.''' motion_bucket_id: NotRequired[int] '''Used by SVD-like models.''' fps: NotRequired[int] '''Used by SVD-like models.''' augmentation_level: NotRequired[float] '''Used by SVD-like models.''' clip_vision_output: NotRequired[ClipVisionOutput_] '''Used by WAN-like models.''' vace_frames: NotRequired[torch.Tensor] '''Used by WAN VACE.''' vace_mask: NotRequired[torch.Tensor] '''Used by WAN VACE.''' vace_strength: NotRequired[float] '''Used by WAN VACE.''' camera_conditions: NotRequired[Any] # TODO: assign proper type once defined '''Used by WAN Camera.''' time_dim_concat: NotRequired[torch.Tensor] '''Used by WAN Phantom Subject.''' time_dim_replace: NotRequired[torch.Tensor] '''Used by Kandinsky5 I2V.''' CondList = list[tuple[torch.Tensor, PooledDict]] Type = CondList @comfytype(io_type="SAMPLER") class Sampler(ComfyTypeIO): if TYPE_CHECKING: Type = Sampler @comfytype(io_type="SIGMAS") class Sigmas(ComfyTypeIO): Type = torch.Tensor @comfytype(io_type="NOISE") class Noise(ComfyTypeIO): Type = torch.Tensor @comfytype(io_type="GUIDER") class Guider(ComfyTypeIO): if TYPE_CHECKING: Type = CFGGuider @comfytype(io_type="CLIP") class Clip(ComfyTypeIO): if TYPE_CHECKING: Type = CLIP @comfytype(io_type="CONTROL_NET") class ControlNet(ComfyTypeIO): if TYPE_CHECKING: Type = ControlNet @comfytype(io_type="VAE") class Vae(ComfyTypeIO): if TYPE_CHECKING: Type = VAE @comfytype(io_type="MODEL") class Model(ComfyTypeIO): if TYPE_CHECKING: Type = ModelPatcher @comfytype(io_type="CLIP_VISION") class ClipVision(ComfyTypeIO): if TYPE_CHECKING: Type = ClipVisionModel @comfytype(io_type="CLIP_VISION_OUTPUT") class ClipVisionOutput(ComfyTypeIO): if TYPE_CHECKING: Type = ClipVisionOutput_ @comfytype(io_type="STYLE_MODEL") class StyleModel(ComfyTypeIO): if TYPE_CHECKING: Type = StyleModel_ @comfytype(io_type="GLIGEN") class Gligen(ComfyTypeIO): '''ModelPatcher that wraps around a 'Gligen' model.''' if TYPE_CHECKING: Type = ModelPatcher @comfytype(io_type="UPSCALE_MODEL") class UpscaleModel(ComfyTypeIO): if TYPE_CHECKING: Type = ImageModelDescriptor @comfytype(io_type="LATENT_UPSCALE_MODEL") class LatentUpscaleModel(ComfyTypeIO): Type = Any @comfytype(io_type="AUDIO") class Audio(ComfyTypeIO): class AudioDict(TypedDict): waveform: torch.Tensor sampler_rate: int Type = AudioDict @comfytype(io_type="VIDEO") class Video(ComfyTypeIO): if TYPE_CHECKING: Type = VideoInput @comfytype(io_type="SVG") class SVG(ComfyTypeIO): Type = _SVG @comfytype(io_type="LORA_MODEL") class LoraModel(ComfyTypeIO): Type = dict[str, torch.Tensor] @comfytype(io_type="LOSS_MAP") class LossMap(ComfyTypeIO): class LossMapDict(TypedDict): loss: list[torch.Tensor] Type = LossMapDict @comfytype(io_type="VOXEL") class Voxel(ComfyTypeIO): Type = VOXEL @comfytype(io_type="MESH") class Mesh(ComfyTypeIO): Type = MESH @comfytype(io_type="FILE_3D") class File3DAny(ComfyTypeIO): """General 3D file type - accepts any supported 3D format.""" Type = File3D @comfytype(io_type="FILE_3D_GLB") class File3DGLB(ComfyTypeIO): """GLB format 3D file - binary glTF, best for web and cross-platform.""" Type = File3D @comfytype(io_type="FILE_3D_GLTF") class File3DGLTF(ComfyTypeIO): """GLTF format 3D file - JSON-based glTF with external resources.""" Type = File3D @comfytype(io_type="FILE_3D_FBX") class File3DFBX(ComfyTypeIO): """FBX format 3D file - best for game engines and animation.""" Type = File3D @comfytype(io_type="FILE_3D_OBJ") class File3DOBJ(ComfyTypeIO): """OBJ format 3D file - simple geometry format.""" Type = File3D @comfytype(io_type="FILE_3D_STL") class File3DSTL(ComfyTypeIO): """STL format 3D file - best for 3D printing.""" Type = File3D @comfytype(io_type="FILE_3D_USDZ") class File3DUSDZ(ComfyTypeIO): """USDZ format 3D file - Apple AR format.""" Type = File3D @comfytype(io_type="HOOKS") class Hooks(ComfyTypeIO): if TYPE_CHECKING: Type = HookGroup @comfytype(io_type="HOOK_KEYFRAMES") class HookKeyframes(ComfyTypeIO): if TYPE_CHECKING: Type = HookKeyframeGroup @comfytype(io_type="TIMESTEPS_RANGE") class TimestepsRange(ComfyTypeIO): '''Range defined by start and endpoint, between 0.0 and 1.0.''' Type = tuple[int, int] @comfytype(io_type="LATENT_OPERATION") class LatentOperation(ComfyTypeIO): Type = Callable[[torch.Tensor], torch.Tensor] @comfytype(io_type="FLOW_CONTROL") class FlowControl(ComfyTypeIO): # NOTE: only used in testing_nodes right now Type = tuple[str, Any] @comfytype(io_type="ACCUMULATION") class Accumulation(ComfyTypeIO): # NOTE: only used in testing_nodes right now class AccumulationDict(TypedDict): accum: list[Any] Type = AccumulationDict @comfytype(io_type="LOAD3D_CAMERA") class Load3DCamera(ComfyTypeIO): class CameraInfo(TypedDict): position: dict[str, float | int] target: dict[str, float | int] zoom: int cameraType: str Type = CameraInfo @comfytype(io_type="LOAD_3D") class Load3D(ComfyTypeIO): """3D models are stored as a dictionary.""" class Model3DDict(TypedDict): image: str mask: str normal: str camera_info: Load3DCamera.CameraInfo recording: NotRequired[str] Type = Model3DDict @comfytype(io_type="LOAD_3D_ANIMATION") class Load3DAnimation(Load3D): ... @comfytype(io_type="PHOTOMAKER") class Photomaker(ComfyTypeIO): Type = Any @comfytype(io_type="POINT") class Point(ComfyTypeIO): Type = Any # NOTE: I couldn't find any references in core code to POINT io_type. Does this exist? @comfytype(io_type="FACE_ANALYSIS") class FaceAnalysis(ComfyTypeIO): Type = Any # NOTE: I couldn't find any references in core code to POINT io_type. Does this exist? @comfytype(io_type="BBOX") class BBOX(ComfyTypeIO): Type = Any # NOTE: I couldn't find any references in core code to POINT io_type. Does this exist? @comfytype(io_type="SEGS") class SEGS(ComfyTypeIO): Type = Any # NOTE: I couldn't find any references in core code to POINT io_type. Does this exist? @comfytype(io_type="*") class AnyType(ComfyTypeIO): Type = Any @comfytype(io_type="MODEL_PATCH") class ModelPatch(ComfyTypeIO): Type = Any @comfytype(io_type="AUDIO_ENCODER") class AudioEncoder(ComfyTypeIO): Type = Any @comfytype(io_type="AUDIO_ENCODER_OUTPUT") class AudioEncoderOutput(ComfyTypeIO): Type = Any @comfytype(io_type="TRACKS") class Tracks(ComfyTypeIO): class TrackDict(TypedDict): track_path: torch.Tensor track_visibility: torch.Tensor Type = TrackDict @comfytype(io_type="COMFY_MULTITYPED_V3") class MultiType: Type = Any class Input(Input): ''' Input that permits more than one input type; if `id` is an instance of `ComfyType.Input`, then that input will be used to create a widget (if applicable) with overridden values. ''' def __init__(self, id: str | Input, types: list[type[_ComfyType] | _ComfyType], display_name: str=None, optional=False, tooltip: str=None, lazy: bool=None, extra_dict=None, raw_link: bool=None, advanced: bool=None): # if id is an Input, then use that Input with overridden values self.input_override = None if isinstance(id, Input): self.input_override = copy.copy(id) optional = id.optional if id.optional is True else optional tooltip = id.tooltip if id.tooltip is not None else tooltip display_name = id.display_name if id.display_name is not None else display_name lazy = id.lazy if id.lazy is not None else lazy id = id.id # if is a widget input, make sure widget_type is set appropriately if isinstance(self.input_override, WidgetInput): self.input_override.widget_type = self.input_override.get_io_type() super().__init__(id, display_name, optional, tooltip, lazy, extra_dict, raw_link, advanced) self._io_types = types @property def io_types(self) -> list[type[Input]]: ''' Returns list of Input class types permitted. ''' io_types = [] for x in self._io_types: if not is_class(x): io_types.append(type(x)) else: io_types.append(x) return io_types def get_io_type(self): # ensure types are unique and order is preserved str_types = [x.io_type for x in self.io_types] if self.input_override is not None: str_types.insert(0, self.input_override.get_io_type()) return ",".join(list(dict.fromkeys(str_types))) def as_dict(self): if self.input_override is not None: return self.input_override.as_dict() | super().as_dict() else: return super().as_dict() @comfytype(io_type="COMFY_MATCHTYPE_V3") class MatchType(ComfyTypeIO): class Template: def __init__(self, template_id: str, allowed_types: _ComfyType | list[_ComfyType] = AnyType): self.template_id = template_id # account for syntactic sugar if not isinstance(allowed_types, Iterable): allowed_types = [allowed_types] for t in allowed_types: if not isinstance(t, type): if not isinstance(t, _ComfyType): raise ValueError(f"Allowed types must be a ComfyType or a list of ComfyTypes, got {t.__class__.__name__}") else: if not issubclass(t, _ComfyType): raise ValueError(f"Allowed types must be a ComfyType or a list of ComfyTypes, got {t.__name__}") self.allowed_types = allowed_types def as_dict(self): return { "template_id": self.template_id, "allowed_types": ",".join([t.io_type for t in self.allowed_types]), } class Input(Input): def __init__(self, id: str, template: MatchType.Template, display_name: str=None, optional=False, tooltip: str=None, lazy: bool=None, extra_dict=None, raw_link: bool=None, advanced: bool=None): super().__init__(id, display_name, optional, tooltip, lazy, extra_dict, raw_link, advanced) self.template = template def as_dict(self): return super().as_dict() | prune_dict({ "template": self.template.as_dict(), }) class Output(Output): def __init__(self, template: MatchType.Template, id: str=None, display_name: str=None, tooltip: str=None, is_output_list=False): if not id and not display_name: display_name = "MATCHTYPE" super().__init__(id, display_name, tooltip, is_output_list) self.template = template def as_dict(self): return super().as_dict() | prune_dict({ "template": self.template.as_dict(), }) class DynamicInput(Input, ABC): ''' Abstract class for dynamic input registration. ''' pass class DynamicOutput(Output, ABC): ''' Abstract class for dynamic output registration. ''' pass def handle_prefix(prefix_list: list[str] | None, id: str | None = None) -> list[str]: if prefix_list is None: prefix_list = [] if id is not None: prefix_list = prefix_list + [id] return prefix_list def finalize_prefix(prefix_list: list[str] | None, id: str | None = None) -> str: assert not (prefix_list is None and id is None) if prefix_list is None: return id elif id is not None: prefix_list = prefix_list + [id] return ".".join(prefix_list) @comfytype(io_type="COMFY_AUTOGROW_V3") class Autogrow(ComfyTypeI): Type = dict[str, Any] _MaxNames = 100 # NOTE: max 100 names for sanity class _AutogrowTemplate: def __init__(self, input: Input): # dynamic inputs are not allowed as the template input assert(not isinstance(input, DynamicInput)) self.input = copy.copy(input) if isinstance(self.input, WidgetInput): self.input.force_input = True self.names: list[str] = [] self.cached_inputs = {} def _create_input(self, input: Input, name: str): new_input = copy.copy(self.input) new_input.id = name return new_input def _create_cached_inputs(self): for name in self.names: self.cached_inputs[name] = self._create_input(self.input, name) def get_all(self) -> list[Input]: return list(self.cached_inputs.values()) def as_dict(self): return prune_dict({ "input": create_input_dict_v1([self.input]), }) def validate(self): self.input.validate() class TemplatePrefix(_AutogrowTemplate): def __init__(self, input: Input, prefix: str, min: int=1, max: int=10): super().__init__(input) self.prefix = prefix assert(min >= 0) assert(max >= 1) assert(max <= Autogrow._MaxNames) self.min = min self.max = max self.names = [f"{self.prefix}{i}" for i in range(self.max)] self._create_cached_inputs() def as_dict(self): return super().as_dict() | prune_dict({ "prefix": self.prefix, "min": self.min, "max": self.max, }) class TemplateNames(_AutogrowTemplate): def __init__(self, input: Input, names: list[str], min: int=1): super().__init__(input) self.names = names[:Autogrow._MaxNames] assert(min >= 0) self.min = min self._create_cached_inputs() def as_dict(self): return super().as_dict() | prune_dict({ "names": self.names, "min": self.min, }) class Input(DynamicInput): def __init__(self, id: str, template: Autogrow.TemplatePrefix | Autogrow.TemplateNames, display_name: str=None, optional=False, tooltip: str=None, lazy: bool=None, extra_dict=None): super().__init__(id, display_name, optional, tooltip, lazy, extra_dict) self.template = template def as_dict(self): return super().as_dict() | prune_dict({ "template": self.template.as_dict(), }) def get_all(self) -> list[Input]: return [self] + self.template.get_all() def validate(self): self.template.validate() @staticmethod def _expand_schema_for_dynamic(out_dict: dict[str, Any], live_inputs: dict[str, Any], value: tuple[str, dict[str, Any]], input_type: str, curr_prefix: list[str] | None): # NOTE: purposely do not include self in out_dict; instead use only the template inputs # need to figure out names based on template type is_names = ("names" in value[1]["template"]) is_prefix = ("prefix" in value[1]["template"]) input = value[1]["template"]["input"] if is_names: min = value[1]["template"]["min"] names = value[1]["template"]["names"] max = len(names) elif is_prefix: prefix = value[1]["template"]["prefix"] min = value[1]["template"]["min"] max = value[1]["template"]["max"] names = [f"{prefix}{i}" for i in range(max)] # need to create a new input based on the contents of input template_input = None template_required = True for _input_type, dict_input in input.items(): # for now, get just the first value from dict_input; if not required, min can be ignored if len(dict_input) == 0: continue template_input = list(dict_input.values())[0] template_required = _input_type == "required" break if template_input is None: raise Exception("template_input could not be determined from required or optional; this should never happen.") new_dict = {} new_dict_added_to = False # first, add possible inputs into out_dict for i, name in enumerate(names): expected_id = finalize_prefix(curr_prefix, name) # required if i < min and template_required: out_dict["required"][expected_id] = template_input type_dict = new_dict.setdefault("required", {}) # optional else: out_dict["optional"][expected_id] = template_input type_dict = new_dict.setdefault("optional", {}) if expected_id in live_inputs: # NOTE: prefix gets added in parse_class_inputs type_dict[name] = template_input new_dict_added_to = True # account for the edge case that all inputs are optional and no values are received if not new_dict_added_to: finalized_prefix = finalize_prefix(curr_prefix) out_dict["dynamic_paths"][finalized_prefix] = finalized_prefix out_dict["dynamic_paths_default_value"][finalized_prefix] = DynamicPathsDefaultValue.EMPTY_DICT parse_class_inputs(out_dict, live_inputs, new_dict, curr_prefix) @comfytype(io_type="COMFY_DYNAMICCOMBO_V3") class DynamicCombo(ComfyTypeI): Type = dict[str, Any] class Option: def __init__(self, key: str, inputs: list[Input]): self.key = key self.inputs = inputs def as_dict(self): return { "key": self.key, "inputs": create_input_dict_v1(self.inputs), } class Input(DynamicInput): def __init__(self, id: str, options: list[DynamicCombo.Option], display_name: str=None, optional=False, tooltip: str=None, lazy: bool=None, extra_dict=None): super().__init__(id, display_name, optional, tooltip, lazy, extra_dict) self.options = options def get_all(self) -> list[Input]: return [self] + [input for option in self.options for input in option.inputs] def as_dict(self): return super().as_dict() | prune_dict({ "options": [o.as_dict() for o in self.options], }) def validate(self): # make sure all nested inputs are validated for option in self.options: for input in option.inputs: input.validate() @staticmethod def _expand_schema_for_dynamic(out_dict: dict[str, Any], live_inputs: dict[str, Any], value: tuple[str, dict[str, Any]], input_type: str, curr_prefix: list[str] | None): finalized_id = finalize_prefix(curr_prefix) if finalized_id in live_inputs: key = live_inputs[finalized_id] selected_option = None # get options from dict options: list[dict[str, str | dict[str, Any]]] = value[1]["options"] for option in options: if option["key"] == key: selected_option = option break if selected_option is not None: parse_class_inputs(out_dict, live_inputs, selected_option["inputs"], curr_prefix) # add self to inputs out_dict[input_type][finalized_id] = value out_dict["dynamic_paths"][finalized_id] = finalize_prefix(curr_prefix, curr_prefix[-1]) @comfytype(io_type="COMFY_DYNAMICSLOT_V3") class DynamicSlot(ComfyTypeI): Type = dict[str, Any] class Input(DynamicInput): def __init__(self, slot: Input, inputs: list[Input], display_name: str=None, tooltip: str=None, lazy: bool=None, extra_dict=None): assert(not isinstance(slot, DynamicInput)) self.slot = copy.copy(slot) self.slot.display_name = slot.display_name if slot.display_name is not None else display_name optional = True self.slot.tooltip = slot.tooltip if slot.tooltip is not None else tooltip self.slot.lazy = slot.lazy if slot.lazy is not None else lazy self.slot.extra_dict = slot.extra_dict if slot.extra_dict is not None else extra_dict super().__init__(slot.id, self.slot.display_name, optional, self.slot.tooltip, self.slot.lazy, self.slot.extra_dict) self.inputs = inputs self.force_input = None # force widget inputs to have no widgets, otherwise this would be awkward if isinstance(self.slot, WidgetInput): self.force_input = True self.slot.force_input = True def get_all(self) -> list[Input]: return [self.slot] + self.inputs def as_dict(self): return super().as_dict() | prune_dict({ "slotType": str(self.slot.get_io_type()), "inputs": create_input_dict_v1(self.inputs), "forceInput": self.force_input, }) def validate(self): self.slot.validate() for input in self.inputs: input.validate() @staticmethod def _expand_schema_for_dynamic(out_dict: dict[str, Any], live_inputs: dict[str, Any], value: tuple[str, dict[str, Any]], input_type: str, curr_prefix: list[str] | None): finalized_id = finalize_prefix(curr_prefix) if finalized_id in live_inputs: inputs = value[1]["inputs"] parse_class_inputs(out_dict, live_inputs, inputs, curr_prefix) # add self to inputs out_dict[input_type][finalized_id] = value out_dict["dynamic_paths"][finalized_id] = finalize_prefix(curr_prefix, curr_prefix[-1]) @comfytype(io_type="IMAGECOMPARE") class ImageCompare(ComfyTypeI): Type = dict class Input(WidgetInput): def __init__(self, id: str, display_name: str=None, optional=False, tooltip: str=None, socketless: bool=True, advanced: bool=None): super().__init__(id, display_name, optional, tooltip, None, None, socketless, None, None, None, None, advanced) def as_dict(self): return super().as_dict() @comfytype(io_type="COLOR") class Color(ComfyTypeIO): Type = str class Input(WidgetInput): def __init__(self, id: str, display_name: str=None, optional=False, tooltip: str=None, socketless: bool=True, advanced: bool=None, default: str="#ffffff"): super().__init__(id, display_name, optional, tooltip, None, default, socketless, None, None, None, None, advanced) self.default: str def as_dict(self): return super().as_dict() @comfytype(io_type="BOUNDING_BOX") class BoundingBox(ComfyTypeIO): class BoundingBoxDict(TypedDict): x: int y: int width: int height: int Type = BoundingBoxDict class Input(WidgetInput): def __init__(self, id: str, display_name: str=None, optional=False, tooltip: str=None, socketless: bool=True, default: dict=None, component: str=None, force_input: bool=None): super().__init__(id, display_name, optional, tooltip, None, default, socketless) self.component = component self.force_input = force_input if default is None: self.default = {"x": 0, "y": 0, "width": 512, "height": 512} def as_dict(self): d = super().as_dict() if self.component: d["component"] = self.component if self.force_input is not None: d["forceInput"] = self.force_input return d DYNAMIC_INPUT_LOOKUP: dict[str, Callable[[dict[str, Any], dict[str, Any], tuple[str, dict[str, Any]], str, list[str] | None], None]] = {} def register_dynamic_input_func(io_type: str, func: Callable[[dict[str, Any], dict[str, Any], tuple[str, dict[str, Any]], str, list[str] | None], None]): DYNAMIC_INPUT_LOOKUP[io_type] = func def get_dynamic_input_func(io_type: str) -> Callable[[dict[str, Any], dict[str, Any], tuple[str, dict[str, Any]], str, list[str] | None], None]: return DYNAMIC_INPUT_LOOKUP[io_type] def setup_dynamic_input_funcs(): # Autogrow.Input register_dynamic_input_func(Autogrow.io_type, Autogrow._expand_schema_for_dynamic) # DynamicCombo.Input register_dynamic_input_func(DynamicCombo.io_type, DynamicCombo._expand_schema_for_dynamic) # DynamicSlot.Input register_dynamic_input_func(DynamicSlot.io_type, DynamicSlot._expand_schema_for_dynamic) if len(DYNAMIC_INPUT_LOOKUP) == 0: setup_dynamic_input_funcs() class V3Data(TypedDict): hidden_inputs: dict[str, Any] 'Dictionary where the keys are the hidden input ids and the values are the values of the hidden inputs.' dynamic_paths: dict[str, Any] 'Dictionary where the keys are the input ids and the values dictate how to turn the inputs into a nested dictionary.' dynamic_paths_default_value: dict[str, Any] 'Dictionary where the keys are the input ids and the values are a string from DynamicPathsDefaultValue for the inputs if value is None.' create_dynamic_tuple: bool 'When True, the value of the dynamic input will be in the format (value, path_key).' class HiddenHolder: def __init__(self, unique_id: str, prompt: Any, extra_pnginfo: Any, dynprompt: Any, auth_token_comfy_org: str, api_key_comfy_org: str, **kwargs): self.unique_id = unique_id """UNIQUE_ID is the unique identifier of the node, and matches the id property of the node on the client side. It is commonly used in client-server communications (see messages).""" self.prompt = prompt """PROMPT is the complete prompt sent by the client to the server. See the prompt object for a full description.""" self.extra_pnginfo = extra_pnginfo """EXTRA_PNGINFO is a dictionary that will be copied into the metadata of any .png files saved. Custom nodes can store additional information in this dictionary for saving (or as a way to communicate with a downstream node).""" self.dynprompt = dynprompt """DYNPROMPT is an instance of comfy_execution.graph.DynamicPrompt. It differs from PROMPT in that it may mutate during the course of execution in response to Node Expansion.""" self.auth_token_comfy_org = auth_token_comfy_org """AUTH_TOKEN_COMFY_ORG is a token acquired from signing into a ComfyOrg account on frontend.""" self.api_key_comfy_org = api_key_comfy_org """API_KEY_COMFY_ORG is an API Key generated by ComfyOrg that allows skipping signing into a ComfyOrg account on frontend.""" def __getattr__(self, key: str): '''If hidden variable not found, return None.''' return None @classmethod def from_dict(cls, d: dict | None): if d is None: d = {} return cls( unique_id=d.get(Hidden.unique_id, None), prompt=d.get(Hidden.prompt, None), extra_pnginfo=d.get(Hidden.extra_pnginfo, None), dynprompt=d.get(Hidden.dynprompt, None), auth_token_comfy_org=d.get(Hidden.auth_token_comfy_org, None), api_key_comfy_org=d.get(Hidden.api_key_comfy_org, None), ) @classmethod def from_v3_data(cls, v3_data: V3Data | None) -> HiddenHolder: return cls.from_dict(v3_data["hidden_inputs"] if v3_data else None) class Hidden(str, Enum): ''' Enumerator for requesting hidden variables in nodes. ''' unique_id = "UNIQUE_ID" """UNIQUE_ID is the unique identifier of the node, and matches the id property of the node on the client side. It is commonly used in client-server communications (see messages).""" prompt = "PROMPT" """PROMPT is the complete prompt sent by the client to the server. See the prompt object for a full description.""" extra_pnginfo = "EXTRA_PNGINFO" """EXTRA_PNGINFO is a dictionary that will be copied into the metadata of any .png files saved. Custom nodes can store additional information in this dictionary for saving (or as a way to communicate with a downstream node).""" dynprompt = "DYNPROMPT" """DYNPROMPT is an instance of comfy_execution.graph.DynamicPrompt. It differs from PROMPT in that it may mutate during the course of execution in response to Node Expansion.""" auth_token_comfy_org = "AUTH_TOKEN_COMFY_ORG" """AUTH_TOKEN_COMFY_ORG is a token acquired from signing into a ComfyOrg account on frontend.""" api_key_comfy_org = "API_KEY_COMFY_ORG" """API_KEY_COMFY_ORG is an API Key generated by ComfyOrg that allows skipping signing into a ComfyOrg account on frontend.""" @dataclass class NodeInfoV1: input: dict=None input_order: dict[str, list[str]]=None is_input_list: bool=None output: list[str]=None output_is_list: list[bool]=None output_name: list[str]=None output_tooltips: list[str]=None output_matchtypes: list[str]=None name: str=None display_name: str=None description: str=None python_module: Any=None category: str=None output_node: bool=None deprecated: bool=None experimental: bool=None dev_only: bool=None api_node: bool=None price_badge: dict | None = None search_aliases: list[str]=None essentials_category: str=None @dataclass class PriceBadgeDepends: widgets: list[str] = field(default_factory=list) inputs: list[str] = field(default_factory=list) input_groups: list[str] = field(default_factory=list) def validate(self) -> None: if not isinstance(self.widgets, list) or any(not isinstance(x, str) for x in self.widgets): raise ValueError("PriceBadgeDepends.widgets must be a list[str].") if not isinstance(self.inputs, list) or any(not isinstance(x, str) for x in self.inputs): raise ValueError("PriceBadgeDepends.inputs must be a list[str].") if not isinstance(self.input_groups, list) or any(not isinstance(x, str) for x in self.input_groups): raise ValueError("PriceBadgeDepends.input_groups must be a list[str].") def as_dict(self, schema_inputs: list["Input"]) -> dict[str, Any]: # Build lookup: widget_id -> io_type input_types: dict[str, str] = {} for inp in schema_inputs: all_inputs = inp.get_all() input_types[inp.id] = inp.get_io_type() # First input is always the parent itself for nested_inp in all_inputs[1:]: # For DynamicCombo/DynamicSlot, nested inputs are prefixed with parent ID # to match frontend naming convention (e.g., "should_texture.enable_pbr") prefixed_id = f"{inp.id}.{nested_inp.id}" input_types[prefixed_id] = nested_inp.get_io_type() # Enrich widgets with type information, raising error for unknown widgets widgets_data: list[dict[str, str]] = [] for w in self.widgets: if w not in input_types: raise ValueError( f"PriceBadge depends_on.widgets references unknown widget '{w}'. " f"Available widgets: {list(input_types.keys())}" ) widgets_data.append({"name": w, "type": input_types[w]}) return { "widgets": widgets_data, "inputs": self.inputs, "input_groups": self.input_groups, } @dataclass class PriceBadge: expr: str depends_on: PriceBadgeDepends = field(default_factory=PriceBadgeDepends) engine: str = field(default="jsonata") def validate(self) -> None: if self.engine != "jsonata": raise ValueError(f"Unsupported PriceBadge.engine '{self.engine}'. Only 'jsonata' is supported.") if not isinstance(self.expr, str) or not self.expr.strip(): raise ValueError("PriceBadge.expr must be a non-empty string.") self.depends_on.validate() def as_dict(self, schema_inputs: list["Input"]) -> dict[str, Any]: return { "engine": self.engine, "depends_on": self.depends_on.as_dict(schema_inputs), "expr": self.expr, } @dataclass class Schema: """Definition of V3 node properties.""" node_id: str """ID of node - should be globally unique. If this is a custom node, add a prefix or postfix to avoid name clashes.""" display_name: str = None """Display name of node.""" category: str = "sd" """The category of the node, as per the "Add Node" menu.""" inputs: list[Input] = field(default_factory=list) outputs: list[Output] = field(default_factory=list) hidden: list[Hidden] = field(default_factory=list) description: str="" """Node description, shown as a tooltip when hovering over the node.""" search_aliases: list[str] = field(default_factory=list) """Alternative names for search. Useful for synonyms, abbreviations, or old names after renaming.""" is_input_list: bool = False """A flag indicating if this node implements the additional code necessary to deal with OUTPUT_IS_LIST nodes. All inputs of ``type`` will become ``list[type]``, regardless of how many items are passed in. This also affects ``check_lazy_status``. From the docs: A node can also override the default input behaviour and receive the whole list in a single call. This is done by setting a class attribute `INPUT_IS_LIST` to ``True``. Comfy Docs: https://docs.comfy.org/custom-nodes/backend/lists#list-processing """ is_output_node: bool=False """Flags this node as an output node, causing any inputs it requires to be executed. If a node is not connected to any output nodes, that node will not be executed. Usage:: From the docs: By default, a node is not considered an output. Set ``OUTPUT_NODE = True`` to specify that it is. Comfy Docs: https://docs.comfy.org/custom-nodes/backend/server_overview#output-node """ is_deprecated: bool=False """Flags a node as deprecated, indicating to users that they should find alternatives to this node.""" is_experimental: bool=False """Flags a node as experimental, informing users that it may change or not work as expected.""" is_dev_only: bool=False """Flags a node as dev-only, hiding it from search/menus unless dev mode is enabled.""" is_api_node: bool=False """Flags a node as an API node. See: https://docs.comfy.org/tutorials/api-nodes/overview.""" price_badge: PriceBadge | None = None """Optional client-evaluated pricing badge declaration for this node.""" not_idempotent: bool=False """Flags a node as not idempotent; when True, the node will run and not reuse the cached outputs when identical inputs are provided on a different node in the graph.""" enable_expand: bool=False """Flags a node as expandable, allowing NodeOutput to include 'expand' property.""" accept_all_inputs: bool=False """When True, all inputs from the prompt will be passed to the node as kwargs, even if not defined in the schema.""" essentials_category: str | None = None """Optional category for the Essentials tab. Path-based like category field (e.g., 'Basic', 'Image Tools/Editing').""" def validate(self): '''Validate the schema: - verify ids on inputs and outputs are unique - both internally and in relation to each other ''' nested_inputs: list[Input] = [] for input in self.inputs: if not isinstance(input, DynamicInput): nested_inputs.extend(input.get_all()) input_ids = [i.id for i in nested_inputs] output_ids = [o.id for o in self.outputs] input_set = set(input_ids) output_set = set(output_ids) issues: list[str] = [] # verify ids are unique per list if len(input_set) != len(input_ids): issues.append(f"Input ids must be unique, but {[item for item, count in Counter(input_ids).items() if count > 1]} are not.") if len(output_set) != len(output_ids): issues.append(f"Output ids must be unique, but {[item for item, count in Counter(output_ids).items() if count > 1]} are not.") if len(issues) > 0: raise ValueError("\n".join(issues)) # validate inputs and outputs for input in self.inputs: input.validate() for output in self.outputs: output.validate() if self.price_badge is not None: self.price_badge.validate() def finalize(self): """Add hidden based on selected schema options, and give outputs without ids default ids.""" # ensure inputs, outputs, and hidden are lists if self.inputs is None: self.inputs = [] if self.outputs is None: self.outputs = [] if self.hidden is None: self.hidden = [] # if is an api_node, will need key-related hidden if self.is_api_node: if Hidden.auth_token_comfy_org not in self.hidden: self.hidden.append(Hidden.auth_token_comfy_org) if Hidden.api_key_comfy_org not in self.hidden: self.hidden.append(Hidden.api_key_comfy_org) # if is an output_node, will need prompt and extra_pnginfo if self.is_output_node: if Hidden.prompt not in self.hidden: self.hidden.append(Hidden.prompt) if Hidden.extra_pnginfo not in self.hidden: self.hidden.append(Hidden.extra_pnginfo) # give outputs without ids default ids for i, output in enumerate(self.outputs): if output.id is None: output.id = f"_{i}_{output.io_type}_" def get_v1_info(self, cls) -> NodeInfoV1: # get V1 inputs input = create_input_dict_v1(self.inputs) if self.hidden: for hidden in self.hidden: input.setdefault("hidden", {})[hidden.name] = (hidden.value,) # create separate lists from output fields output = [] output_is_list = [] output_name = [] output_tooltips = [] output_matchtypes = [] any_matchtypes = False if self.outputs: for o in self.outputs: output.append(o.io_type) output_is_list.append(o.is_output_list) output_name.append(o.display_name if o.display_name else o.io_type) output_tooltips.append(o.tooltip if o.tooltip else None) # special handling for MatchType if isinstance(o, MatchType.Output): output_matchtypes.append(o.template.template_id) any_matchtypes = True else: output_matchtypes.append(None) # clear out lists that are all None if not any_matchtypes: output_matchtypes = None info = NodeInfoV1( input=input, input_order={key: list(value.keys()) for (key, value) in input.items()}, is_input_list=self.is_input_list, output=output, output_is_list=output_is_list, output_name=output_name, output_tooltips=output_tooltips, output_matchtypes=output_matchtypes, name=self.node_id, display_name=self.display_name, category=self.category, description=self.description, output_node=self.is_output_node, deprecated=self.is_deprecated, experimental=self.is_experimental, dev_only=self.is_dev_only, api_node=self.is_api_node, python_module=getattr(cls, "RELATIVE_PYTHON_MODULE", "nodes"), price_badge=self.price_badge.as_dict(self.inputs) if self.price_badge is not None else None, search_aliases=self.search_aliases if self.search_aliases else None, essentials_category=self.essentials_category, ) return info def get_finalized_class_inputs(d: dict[str, Any], live_inputs: dict[str, Any], include_hidden=False) -> tuple[dict[str, Any], V3Data]: out_dict = { "required": {}, "optional": {}, "dynamic_paths": {}, "dynamic_paths_default_value": {}, } d = d.copy() # ignore hidden for parsing hidden = d.pop("hidden", None) parse_class_inputs(out_dict, live_inputs, d) if hidden is not None and include_hidden: out_dict["hidden"] = hidden v3_data = {} dynamic_paths = out_dict.pop("dynamic_paths", None) if dynamic_paths is not None and len(dynamic_paths) > 0: v3_data["dynamic_paths"] = dynamic_paths # this list is used for autogrow, in the case all inputs are optional and no values are passed dynamic_paths_default_value = out_dict.pop("dynamic_paths_default_value", None) if dynamic_paths_default_value is not None and len(dynamic_paths_default_value) > 0: v3_data["dynamic_paths_default_value"] = dynamic_paths_default_value return out_dict, hidden, v3_data def parse_class_inputs(out_dict: dict[str, Any], live_inputs: dict[str, Any], curr_dict: dict[str, Any], curr_prefix: list[str] | None=None) -> None: for input_type, inner_d in curr_dict.items(): for id, value in inner_d.items(): io_type = value[0] if io_type in DYNAMIC_INPUT_LOOKUP: # dynamic inputs need to be handled with lookup functions dynamic_input_func = get_dynamic_input_func(io_type) new_prefix = handle_prefix(curr_prefix, id) dynamic_input_func(out_dict, live_inputs, value, input_type, new_prefix) else: # non-dynamic inputs get directly transferred finalized_id = finalize_prefix(curr_prefix, id) out_dict[input_type][finalized_id] = value if curr_prefix: out_dict["dynamic_paths"][finalized_id] = finalized_id def create_input_dict_v1(inputs: list[Input]) -> dict: input = { "required": {} } for i in inputs: add_to_dict_v1(i, input) return input def add_to_dict_v1(i: Input, d: dict): key = "optional" if i.optional else "required" as_dict = i.as_dict() # for v1, we don't want to include the optional key as_dict.pop("optional", None) d.setdefault(key, {})[i.id] = (i.get_io_type(), as_dict) class DynamicPathsDefaultValue: EMPTY_DICT = "empty_dict" def build_nested_inputs(values: dict[str, Any], v3_data: V3Data): paths = v3_data.get("dynamic_paths", None) default_value_dict = v3_data.get("dynamic_paths_default_value", {}) if paths is None: return values values = values.copy() result = {} create_tuple = v3_data.get("create_dynamic_tuple", False) for key, path in paths.items(): parts = path.split(".") current = result for i, p in enumerate(parts): is_last = (i == len(parts) - 1) if is_last: value = values.pop(key, None) if value is None: # see if a default value was provided for this key default_option = default_value_dict.get(key, None) if default_option == DynamicPathsDefaultValue.EMPTY_DICT: value = {} if create_tuple: value = (value, key) current[p] = value else: current = current.setdefault(p, {}) values.update(result) return values class _ComfyNodeBaseInternal(_ComfyNodeInternal): """Common base class for storing internal methods and properties; DO NOT USE for defining nodes.""" RELATIVE_PYTHON_MODULE = None SCHEMA = None # filled in during execution hidden: HiddenHolder = None @classmethod @abstractmethod def define_schema(cls) -> Schema: """Override this function with one that returns a Schema instance.""" raise NotImplementedError @classmethod @abstractmethod def execute(cls, **kwargs) -> NodeOutput: """Override this function with one that performs node's actions.""" raise NotImplementedError @classmethod def validate_inputs(cls, **kwargs) -> bool | str: """Optionally, define this function to validate inputs; equivalent to V1's VALIDATE_INPUTS. If the function returns a string, it will be used as the validation error message for the node. """ raise NotImplementedError @classmethod def fingerprint_inputs(cls, **kwargs) -> Any: """Optionally, define this function to fingerprint inputs; equivalent to V1's IS_CHANGED. If this function returns the same value as last run, the node will not be executed.""" raise NotImplementedError @classmethod def check_lazy_status(cls, **kwargs) -> list[str]: """Optionally, define this function to return a list of input names that should be evaluated. This basic mixin impl. requires all inputs. :kwargs: All node inputs will be included here. If the input is ``None``, it should be assumed that it has not yet been evaluated. \ When using ``INPUT_IS_LIST = True``, unevaluated will instead be ``(None,)``. Params should match the nodes execution ``FUNCTION`` (self, and all inputs by name). Will be executed repeatedly until it returns an empty list, or all requested items were already evaluated (and sent as params). Comfy Docs: https://docs.comfy.org/custom-nodes/backend/lazy_evaluation#defining-check-lazy-status """ return [name for name in kwargs if kwargs[name] is None] def __init__(self): self.__class__.VALIDATE_CLASS() @classmethod def GET_BASE_CLASS(cls): return _ComfyNodeBaseInternal @final @classmethod def VALIDATE_CLASS(cls): if first_real_override(cls, "define_schema") is None: raise Exception(f"No define_schema function was defined for node class {cls.__name__}.") if first_real_override(cls, "execute") is None: raise Exception(f"No execute function was defined for node class {cls.__name__}.") @classproperty def FUNCTION(cls): # noqa if inspect.iscoroutinefunction(cls.execute): return "EXECUTE_NORMALIZED_ASYNC" return "EXECUTE_NORMALIZED" @final @classmethod def EXECUTE_NORMALIZED(cls, *args, **kwargs) -> NodeOutput: to_return = cls.execute(*args, **kwargs) if to_return is None: to_return = NodeOutput() elif isinstance(to_return, NodeOutput): pass elif isinstance(to_return, tuple): to_return = NodeOutput(*to_return) elif isinstance(to_return, dict): to_return = NodeOutput.from_dict(to_return) elif isinstance(to_return, ExecutionBlocker): to_return = NodeOutput(block_execution=to_return.message) else: raise Exception(f"Invalid return type from node: {type(to_return)}") if to_return.expand is not None and not cls.SCHEMA.enable_expand: raise Exception(f"Node {cls.__name__} is not expandable, but expand included in NodeOutput; developer should set enable_expand=True on node's Schema to allow this.") return to_return @final @classmethod async def EXECUTE_NORMALIZED_ASYNC(cls, *args, **kwargs) -> NodeOutput: to_return = await cls.execute(*args, **kwargs) if to_return is None: to_return = NodeOutput() elif isinstance(to_return, NodeOutput): pass elif isinstance(to_return, tuple): to_return = NodeOutput(*to_return) elif isinstance(to_return, dict): to_return = NodeOutput.from_dict(to_return) elif isinstance(to_return, ExecutionBlocker): to_return = NodeOutput(block_execution=to_return.message) else: raise Exception(f"Invalid return type from node: {type(to_return)}") if to_return.expand is not None and not cls.SCHEMA.enable_expand: raise Exception(f"Node {cls.__name__} is not expandable, but expand included in NodeOutput; developer should set enable_expand=True on node's Schema to allow this.") return to_return @final @classmethod def PREPARE_CLASS_CLONE(cls, v3_data: V3Data | None) -> type[ComfyNode]: """Creates clone of real node class to prevent monkey-patching.""" c_type: type[ComfyNode] = cls if is_class(cls) else type(cls) type_clone: type[ComfyNode] = shallow_clone_class(c_type) # set hidden type_clone.hidden = HiddenHolder.from_v3_data(v3_data) return type_clone ############################################# # V1 Backwards Compatibility code #-------------------------------------------- @final @classmethod def GET_NODE_INFO_V1(cls) -> dict[str, Any]: schema = cls.GET_SCHEMA() info = schema.get_v1_info(cls) return asdict(info) _DESCRIPTION = None @final @classproperty def DESCRIPTION(cls): # noqa if cls._DESCRIPTION is None: cls.GET_SCHEMA() return cls._DESCRIPTION _CATEGORY = None @final @classproperty def CATEGORY(cls): # noqa if cls._CATEGORY is None: cls.GET_SCHEMA() return cls._CATEGORY _EXPERIMENTAL = None @final @classproperty def EXPERIMENTAL(cls): # noqa if cls._EXPERIMENTAL is None: cls.GET_SCHEMA() return cls._EXPERIMENTAL _DEPRECATED = None @final @classproperty def DEPRECATED(cls): # noqa if cls._DEPRECATED is None: cls.GET_SCHEMA() return cls._DEPRECATED _DEV_ONLY = None @final @classproperty def DEV_ONLY(cls): # noqa if cls._DEV_ONLY is None: cls.GET_SCHEMA() return cls._DEV_ONLY _API_NODE = None @final @classproperty def API_NODE(cls): # noqa if cls._API_NODE is None: cls.GET_SCHEMA() return cls._API_NODE _OUTPUT_NODE = None @final @classproperty def OUTPUT_NODE(cls): # noqa if cls._OUTPUT_NODE is None: cls.GET_SCHEMA() return cls._OUTPUT_NODE _INPUT_IS_LIST = None @final @classproperty def INPUT_IS_LIST(cls): # noqa if cls._INPUT_IS_LIST is None: cls.GET_SCHEMA() return cls._INPUT_IS_LIST _OUTPUT_IS_LIST = None @final @classproperty def OUTPUT_IS_LIST(cls): # noqa if cls._OUTPUT_IS_LIST is None: cls.GET_SCHEMA() return cls._OUTPUT_IS_LIST _RETURN_TYPES = None @final @classproperty def RETURN_TYPES(cls): # noqa if cls._RETURN_TYPES is None: cls.GET_SCHEMA() return cls._RETURN_TYPES _RETURN_NAMES = None @final @classproperty def RETURN_NAMES(cls): # noqa if cls._RETURN_NAMES is None: cls.GET_SCHEMA() return cls._RETURN_NAMES _OUTPUT_TOOLTIPS = None @final @classproperty def OUTPUT_TOOLTIPS(cls): # noqa if cls._OUTPUT_TOOLTIPS is None: cls.GET_SCHEMA() return cls._OUTPUT_TOOLTIPS _NOT_IDEMPOTENT = None @final @classproperty def NOT_IDEMPOTENT(cls): # noqa if cls._NOT_IDEMPOTENT is None: cls.GET_SCHEMA() return cls._NOT_IDEMPOTENT _ACCEPT_ALL_INPUTS = None @final @classproperty def ACCEPT_ALL_INPUTS(cls): # noqa if cls._ACCEPT_ALL_INPUTS is None: cls.GET_SCHEMA() return cls._ACCEPT_ALL_INPUTS @final @classmethod def INPUT_TYPES(cls) -> dict[str, dict]: schema = cls.FINALIZE_SCHEMA() info = schema.get_v1_info(cls) return info.input @final @classmethod def FINALIZE_SCHEMA(cls): """Call define_schema and finalize it.""" schema = cls.define_schema() schema.finalize() return schema @final @classmethod def GET_SCHEMA(cls) -> Schema: """Validate node class, finalize schema, validate schema, and set expected class properties.""" cls.VALIDATE_CLASS() schema = cls.FINALIZE_SCHEMA() schema.validate() if cls._DESCRIPTION is None: cls._DESCRIPTION = schema.description if cls._CATEGORY is None: cls._CATEGORY = schema.category if cls._EXPERIMENTAL is None: cls._EXPERIMENTAL = schema.is_experimental if cls._DEPRECATED is None: cls._DEPRECATED = schema.is_deprecated if cls._DEV_ONLY is None: cls._DEV_ONLY = schema.is_dev_only if cls._API_NODE is None: cls._API_NODE = schema.is_api_node if cls._OUTPUT_NODE is None: cls._OUTPUT_NODE = schema.is_output_node if cls._INPUT_IS_LIST is None: cls._INPUT_IS_LIST = schema.is_input_list if cls._NOT_IDEMPOTENT is None: cls._NOT_IDEMPOTENT = schema.not_idempotent if cls._ACCEPT_ALL_INPUTS is None: cls._ACCEPT_ALL_INPUTS = schema.accept_all_inputs if cls._RETURN_TYPES is None: output = [] output_name = [] output_is_list = [] output_tooltips = [] if schema.outputs: for o in schema.outputs: output.append(o.io_type) output_name.append(o.display_name if o.display_name else o.io_type) output_is_list.append(o.is_output_list) output_tooltips.append(o.tooltip if o.tooltip else None) cls._RETURN_TYPES = output cls._RETURN_NAMES = output_name cls._OUTPUT_IS_LIST = output_is_list cls._OUTPUT_TOOLTIPS = output_tooltips cls.SCHEMA = schema return schema #-------------------------------------------- ############################################# class ComfyNode(_ComfyNodeBaseInternal): """Common base class for all V3 nodes.""" @classmethod @abstractmethod def define_schema(cls) -> Schema: """Override this function with one that returns a Schema instance.""" raise NotImplementedError @classmethod @abstractmethod def execute(cls, **kwargs) -> NodeOutput: """Override this function with one that performs node's actions.""" raise NotImplementedError @classmethod def validate_inputs(cls, **kwargs) -> bool | str: """Optionally, define this function to validate inputs; equivalent to V1's VALIDATE_INPUTS.""" raise NotImplementedError @classmethod def fingerprint_inputs(cls, **kwargs) -> Any: """Optionally, define this function to fingerprint inputs; equivalent to V1's IS_CHANGED.""" raise NotImplementedError @classmethod def check_lazy_status(cls, **kwargs) -> list[str]: """Optionally, define this function to return a list of input names that should be evaluated. This basic mixin impl. requires all inputs. :kwargs: All node inputs will be included here. If the input is ``None``, it should be assumed that it has not yet been evaluated. \ When using ``INPUT_IS_LIST = True``, unevaluated will instead be ``(None,)``. Params should match the nodes execution ``FUNCTION`` (self, and all inputs by name). Will be executed repeatedly until it returns an empty list, or all requested items were already evaluated (and sent as params). Comfy Docs: https://docs.comfy.org/custom-nodes/backend/lazy_evaluation#defining-check-lazy-status """ return [name for name in kwargs if kwargs[name] is None] @final @classmethod def GET_BASE_CLASS(cls): """DO NOT override this class. Will break things in execution.py.""" return ComfyNode class NodeOutput(_NodeOutputInternal): ''' Standardized output of a node; can pass in any number of args and/or a UIOutput into 'ui' kwarg. ''' def __init__(self, *args: Any, ui: _UIOutput | dict=None, expand: dict=None, block_execution: str=None): self.args = args self.ui = ui self.expand = expand self.block_execution = block_execution @property def result(self): return self.args if len(self.args) > 0 else None @classmethod def from_dict(cls, data: dict[str, Any]) -> NodeOutput: args = () ui = None expand = None if "result" in data: result = data["result"] if isinstance(result, ExecutionBlocker): return cls(block_execution=result.message) args = result if "ui" in data: ui = data["ui"] if "expand" in data: expand = data["expand"] return cls(*args, ui=ui, expand=expand) def __getitem__(self, index) -> Any: return self.args[index] class _UIOutput(ABC): def __init__(self): pass @abstractmethod def as_dict(self) -> dict: ... class InputMapOldId(TypedDict): """Map an old node input to a new node input by ID.""" new_id: str old_id: str class InputMapSetValue(TypedDict): """Set a specific value for a new node input.""" new_id: str set_value: Any InputMap = InputMapOldId | InputMapSetValue """ Input mapping for node replacement. Type is inferred by dictionary keys: - {"new_id": str, "old_id": str} - maps old input to new input - {"new_id": str, "set_value": Any} - sets a specific value for new input """ class OutputMap(TypedDict): """Map outputs of node replacement via indexes.""" new_idx: int old_idx: int class NodeReplace: """ Defines a possible node replacement, mapping inputs and outputs of the old node to the new node. Also supports assigning specific values to the input widgets of the new node. Args: new_node_id: The class name of the new replacement node. old_node_id: The class name of the deprecated node. old_widget_ids: Ordered list of input IDs for widgets that may not have an input slot connected. The workflow JSON stores widget values by their relative position index, not by ID. This list maps those positional indexes to input IDs, enabling the replacement system to correctly identify widget values during node migration. input_mapping: List of input mappings from old node to new node. output_mapping: List of output mappings from old node to new node. """ def __init__(self, new_node_id: str, old_node_id: str, old_widget_ids: list[str] | None=None, input_mapping: list[InputMap] | None=None, output_mapping: list[OutputMap] | None=None, ): self.new_node_id = new_node_id self.old_node_id = old_node_id self.old_widget_ids = old_widget_ids self.input_mapping = input_mapping self.output_mapping = output_mapping def as_dict(self): """Create serializable representation of the node replacement.""" return { "new_node_id": self.new_node_id, "old_node_id": self.old_node_id, "old_widget_ids": self.old_widget_ids, "input_mapping": list(self.input_mapping) if self.input_mapping else None, "output_mapping": list(self.output_mapping) if self.output_mapping else None, } __all__ = [ "FolderType", "UploadType", "RemoteOptions", "NumberDisplay", "ControlAfterGenerate", "comfytype", "Custom", "Input", "WidgetInput", "Output", "ComfyTypeI", "ComfyTypeIO", # Supported Types "Boolean", "Int", "Float", "String", "Combo", "MultiCombo", "Image", "WanCameraEmbedding", "Webcam", "Mask", "Latent", "Conditioning", "Sampler", "Sigmas", "Noise", "Guider", "Clip", "ControlNet", "Vae", "Model", "ModelPatch", "ClipVision", "ClipVisionOutput", "AudioEncoder", "AudioEncoderOutput", "StyleModel", "Gligen", "UpscaleModel", "LatentUpscaleModel", "Audio", "Video", "SVG", "LoraModel", "LossMap", "Voxel", "Mesh", "File3DAny", "File3DGLB", "File3DGLTF", "File3DFBX", "File3DOBJ", "File3DSTL", "File3DUSDZ", "Hooks", "HookKeyframes", "TimestepsRange", "LatentOperation", "FlowControl", "Accumulation", "Load3DCamera", "Load3D", "Load3DAnimation", "Photomaker", "Point", "FaceAnalysis", "BBOX", "SEGS", "AnyType", "MultiType", "Tracks", "Color", # Dynamic Types "MatchType", "DynamicCombo", "Autogrow", # Other classes "HiddenHolder", "Hidden", "NodeInfoV1", "Schema", "ComfyNode", "NodeOutput", "add_to_dict_v1", "V3Data", "ImageCompare", "PriceBadgeDepends", "PriceBadge", "BoundingBox", "NodeReplace", ]
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api/latest/_io.py", "license": "GNU General Public License v3.0", "lines": 1921, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_api/latest/_ui.py
from __future__ import annotations import json import os import random import uuid from io import BytesIO import av import numpy as np import torch try: import torchaudio TORCH_AUDIO_AVAILABLE = True except: TORCH_AUDIO_AVAILABLE = False from PIL import Image as PILImage from PIL.PngImagePlugin import PngInfo import folder_paths # used for image preview from comfy.cli_args import args from ._io import ComfyNode, FolderType, Image, _UIOutput class SavedResult(dict): def __init__(self, filename: str, subfolder: str, type: FolderType): super().__init__(filename=filename, subfolder=subfolder,type=type.value) @property def filename(self) -> str: return self["filename"] @property def subfolder(self) -> str: return self["subfolder"] @property def type(self) -> FolderType: return FolderType(self["type"]) class SavedImages(_UIOutput): """A UI output class to represent one or more saved images, potentially animated.""" def __init__(self, results: list[SavedResult], is_animated: bool = False): super().__init__() self.results = results self.is_animated = is_animated def as_dict(self) -> dict: data = {"images": self.results} if self.is_animated: data["animated"] = (True,) return data class SavedAudios(_UIOutput): """UI wrapper around one or more audio files on disk (FLAC / MP3 / Opus).""" def __init__(self, results: list[SavedResult]): super().__init__() self.results = results def as_dict(self) -> dict: return {"audio": self.results} def _get_directory_by_folder_type(folder_type: FolderType) -> str: if folder_type == FolderType.input: return folder_paths.get_input_directory() if folder_type == FolderType.output: return folder_paths.get_output_directory() return folder_paths.get_temp_directory() class ImageSaveHelper: """A helper class with static methods to handle image saving and metadata.""" @staticmethod def _convert_tensor_to_pil(image_tensor: torch.Tensor) -> PILImage.Image: """Converts a single torch tensor to a PIL Image.""" return PILImage.fromarray(np.clip(255.0 * image_tensor.cpu().numpy(), 0, 255).astype(np.uint8)) @staticmethod def _create_png_metadata(cls: type[ComfyNode] | None) -> PngInfo | None: """Creates a PngInfo object with prompt and extra_pnginfo.""" if args.disable_metadata or cls is None or not cls.hidden: return None metadata = PngInfo() if cls.hidden.prompt: metadata.add_text("prompt", json.dumps(cls.hidden.prompt)) if cls.hidden.extra_pnginfo: for x in cls.hidden.extra_pnginfo: metadata.add_text(x, json.dumps(cls.hidden.extra_pnginfo[x])) return metadata @staticmethod def _create_animated_png_metadata(cls: type[ComfyNode] | None) -> PngInfo | None: """Creates a PngInfo object with prompt and extra_pnginfo for animated PNGs (APNG).""" if args.disable_metadata or cls is None or not cls.hidden: return None metadata = PngInfo() if cls.hidden.prompt: metadata.add( b"comf", "prompt".encode("latin-1", "strict") + b"\0" + json.dumps(cls.hidden.prompt).encode("latin-1", "strict"), after_idat=True, ) if cls.hidden.extra_pnginfo: for x in cls.hidden.extra_pnginfo: metadata.add( b"comf", x.encode("latin-1", "strict") + b"\0" + json.dumps(cls.hidden.extra_pnginfo[x]).encode("latin-1", "strict"), after_idat=True, ) return metadata @staticmethod def _create_webp_metadata(pil_image: PILImage.Image, cls: type[ComfyNode] | None) -> PILImage.Exif: """Creates EXIF metadata bytes for WebP images.""" exif_data = pil_image.getexif() if args.disable_metadata or cls is None or cls.hidden is None: return exif_data if cls.hidden.prompt is not None: exif_data[0x0110] = "prompt:{}".format(json.dumps(cls.hidden.prompt)) # EXIF 0x0110 = Model if cls.hidden.extra_pnginfo is not None: inital_exif_tag = 0x010F # EXIF 0x010f = Make for key, value in cls.hidden.extra_pnginfo.items(): exif_data[inital_exif_tag] = "{}:{}".format(key, json.dumps(value)) inital_exif_tag -= 1 return exif_data @staticmethod def save_images( images, filename_prefix: str, folder_type: FolderType, cls: type[ComfyNode] | None, compress_level = 4, ) -> list[SavedResult]: """Saves a batch of images as individual PNG files.""" full_output_folder, filename, counter, subfolder, _ = folder_paths.get_save_image_path( filename_prefix, _get_directory_by_folder_type(folder_type), images[0].shape[1], images[0].shape[0] ) results = [] metadata = ImageSaveHelper._create_png_metadata(cls) for batch_number, image_tensor in enumerate(images): img = ImageSaveHelper._convert_tensor_to_pil(image_tensor) filename_with_batch_num = filename.replace("%batch_num%", str(batch_number)) file = f"{filename_with_batch_num}_{counter:05}_.png" img.save(os.path.join(full_output_folder, file), pnginfo=metadata, compress_level=compress_level) results.append(SavedResult(file, subfolder, folder_type)) counter += 1 return results @staticmethod def get_save_images_ui(images, filename_prefix: str, cls: type[ComfyNode] | None, compress_level=4) -> SavedImages: """Saves a batch of images and returns a UI object for the node output.""" return SavedImages( ImageSaveHelper.save_images( images, filename_prefix=filename_prefix, folder_type=FolderType.output, cls=cls, compress_level=compress_level, ) ) @staticmethod def save_animated_png( images, filename_prefix: str, folder_type: FolderType, cls: type[ComfyNode] | None, fps: float, compress_level: int ) -> SavedResult: """Saves a batch of images as a single animated PNG.""" full_output_folder, filename, counter, subfolder, _ = folder_paths.get_save_image_path( filename_prefix, _get_directory_by_folder_type(folder_type), images[0].shape[1], images[0].shape[0] ) pil_images = [ImageSaveHelper._convert_tensor_to_pil(img) for img in images] metadata = ImageSaveHelper._create_animated_png_metadata(cls) file = f"{filename}_{counter:05}_.png" save_path = os.path.join(full_output_folder, file) pil_images[0].save( save_path, pnginfo=metadata, compress_level=compress_level, save_all=True, duration=int(1000.0 / fps), append_images=pil_images[1:], ) return SavedResult(file, subfolder, folder_type) @staticmethod def get_save_animated_png_ui( images, filename_prefix: str, cls: type[ComfyNode] | None, fps: float, compress_level: int ) -> SavedImages: """Saves an animated PNG and returns a UI object for the node output.""" result = ImageSaveHelper.save_animated_png( images, filename_prefix=filename_prefix, folder_type=FolderType.output, cls=cls, fps=fps, compress_level=compress_level, ) return SavedImages([result], is_animated=len(images) > 1) @staticmethod def save_animated_webp( images, filename_prefix: str, folder_type: FolderType, cls: type[ComfyNode] | None, fps: float, lossless: bool, quality: int, method: int, ) -> SavedResult: """Saves a batch of images as a single animated WebP.""" full_output_folder, filename, counter, subfolder, _ = folder_paths.get_save_image_path( filename_prefix, _get_directory_by_folder_type(folder_type), images[0].shape[1], images[0].shape[0] ) pil_images = [ImageSaveHelper._convert_tensor_to_pil(img) for img in images] pil_exif = ImageSaveHelper._create_webp_metadata(pil_images[0], cls) file = f"{filename}_{counter:05}_.webp" pil_images[0].save( os.path.join(full_output_folder, file), save_all=True, duration=int(1000.0 / fps), append_images=pil_images[1:], exif=pil_exif, lossless=lossless, quality=quality, method=method, ) return SavedResult(file, subfolder, folder_type) @staticmethod def get_save_animated_webp_ui( images, filename_prefix: str, cls: type[ComfyNode] | None, fps: float, lossless: bool, quality: int, method: int, ) -> SavedImages: """Saves an animated WebP and returns a UI object for the node output.""" result = ImageSaveHelper.save_animated_webp( images, filename_prefix=filename_prefix, folder_type=FolderType.output, cls=cls, fps=fps, lossless=lossless, quality=quality, method=method, ) return SavedImages([result], is_animated=len(images) > 1) class AudioSaveHelper: """A helper class with static methods to handle audio saving and metadata.""" _OPUS_RATES = [8000, 12000, 16000, 24000, 48000] @staticmethod def save_audio( audio: dict, filename_prefix: str, folder_type: FolderType, cls: type[ComfyNode] | None, format: str = "flac", quality: str = "128k", ) -> list[SavedResult]: full_output_folder, filename, counter, subfolder, _ = folder_paths.get_save_image_path( filename_prefix, _get_directory_by_folder_type(folder_type) ) metadata = {} if not args.disable_metadata and cls is not None: if cls.hidden.prompt is not None: metadata["prompt"] = json.dumps(cls.hidden.prompt) if cls.hidden.extra_pnginfo is not None: for x in cls.hidden.extra_pnginfo: metadata[x] = json.dumps(cls.hidden.extra_pnginfo[x]) results = [] for batch_number, waveform in enumerate(audio["waveform"].cpu()): filename_with_batch_num = filename.replace("%batch_num%", str(batch_number)) file = f"{filename_with_batch_num}_{counter:05}_.{format}" output_path = os.path.join(full_output_folder, file) # Use original sample rate initially sample_rate = audio["sample_rate"] # Handle Opus sample rate requirements if format == "opus": if sample_rate > 48000: sample_rate = 48000 elif sample_rate not in AudioSaveHelper._OPUS_RATES: # Find the next highest supported rate for rate in sorted(AudioSaveHelper._OPUS_RATES): if rate > sample_rate: sample_rate = rate break if sample_rate not in AudioSaveHelper._OPUS_RATES: # Fallback if still not supported sample_rate = 48000 # Resample if necessary if sample_rate != audio["sample_rate"]: if not TORCH_AUDIO_AVAILABLE: raise Exception("torchaudio is not available; cannot resample audio.") waveform = torchaudio.functional.resample(waveform, audio["sample_rate"], sample_rate) # Create output with specified format output_buffer = BytesIO() output_container = av.open(output_buffer, mode="w", format=format) # Set metadata on the container for key, value in metadata.items(): output_container.metadata[key] = value layout = "mono" if waveform.shape[0] == 1 else "stereo" # Set up the output stream with appropriate properties if format == "opus": out_stream = output_container.add_stream("libopus", rate=sample_rate, layout=layout) if quality == "64k": out_stream.bit_rate = 64000 elif quality == "96k": out_stream.bit_rate = 96000 elif quality == "128k": out_stream.bit_rate = 128000 elif quality == "192k": out_stream.bit_rate = 192000 elif quality == "320k": out_stream.bit_rate = 320000 elif format == "mp3": out_stream = output_container.add_stream("libmp3lame", rate=sample_rate, layout=layout) if quality == "V0": # TODO i would really love to support V3 and V5 but there doesn't seem to be a way to set the qscale level, the property below is a bool out_stream.codec_context.qscale = 1 elif quality == "128k": out_stream.bit_rate = 128000 elif quality == "320k": out_stream.bit_rate = 320000 else: # format == "flac": out_stream = output_container.add_stream("flac", rate=sample_rate, layout=layout) frame = av.AudioFrame.from_ndarray( waveform.movedim(0, 1).reshape(1, -1).float().numpy(), format="flt", layout=layout, ) frame.sample_rate = sample_rate frame.pts = 0 output_container.mux(out_stream.encode(frame)) # Flush encoder output_container.mux(out_stream.encode(None)) # Close containers output_container.close() # Write the output to file output_buffer.seek(0) with open(output_path, "wb") as f: f.write(output_buffer.getbuffer()) results.append(SavedResult(file, subfolder, folder_type)) counter += 1 return results @staticmethod def get_save_audio_ui( audio, filename_prefix: str, cls: type[ComfyNode] | None, format: str = "flac", quality: str = "128k", ) -> SavedAudios: """Save and instantly wrap for UI.""" return SavedAudios( AudioSaveHelper.save_audio( audio, filename_prefix=filename_prefix, folder_type=FolderType.output, cls=cls, format=format, quality=quality, ) ) class PreviewImage(_UIOutput): def __init__(self, image: Image.Type, animated: bool = False, cls: type[ComfyNode] = None, **kwargs): self.values = ImageSaveHelper.save_images( image, filename_prefix="ComfyUI_temp_" + ''.join(random.choice("abcdefghijklmnopqrstupvxyz") for _ in range(5)), folder_type=FolderType.temp, cls=cls, compress_level=1, ) self.animated = animated def as_dict(self): return { "images": self.values, "animated": (self.animated,) } class PreviewMask(PreviewImage): def __init__(self, mask: PreviewMask.Type, animated: bool=False, cls: ComfyNode=None, **kwargs): preview = mask.reshape((-1, 1, mask.shape[-2], mask.shape[-1])).movedim(1, -1).expand(-1, -1, -1, 3) super().__init__(preview, animated, cls, **kwargs) class PreviewAudio(_UIOutput): def __init__(self, audio: dict, cls: type[ComfyNode] = None, **kwargs): self.values = AudioSaveHelper.save_audio( audio, filename_prefix="ComfyUI_temp_" + "".join(random.choice("abcdefghijklmnopqrstuvwxyz") for _ in range(5)), folder_type=FolderType.temp, cls=cls, format="flac", quality="128k", ) def as_dict(self) -> dict: return {"audio": self.values} class PreviewVideo(_UIOutput): def __init__(self, values: list[SavedResult | dict], **kwargs): self.values = values def as_dict(self): return {"images": self.values, "animated": (True,)} class PreviewUI3D(_UIOutput): def __init__(self, model_file, camera_info, **kwargs): self.model_file = model_file self.camera_info = camera_info self.bg_image_path = None bg_image = kwargs.get("bg_image", None) if bg_image is not None: img_array = (bg_image[0].cpu().numpy() * 255).astype(np.uint8) img = PILImage.fromarray(img_array) temp_dir = folder_paths.get_temp_directory() filename = f"bg_{uuid.uuid4().hex}.png" bg_image_path = os.path.join(temp_dir, filename) img.save(bg_image_path, compress_level=1) self.bg_image_path = f"temp/{filename}" def as_dict(self): return {"result": [self.model_file, self.camera_info, self.bg_image_path]} class PreviewText(_UIOutput): def __init__(self, value: str, **kwargs): self.value = value def as_dict(self): return {"text": (self.value,)} __all__ = [ "SavedResult", "SavedImages", "SavedAudios", "ImageSaveHelper", "AudioSaveHelper", "PreviewImage", "PreviewMask", "PreviewAudio", "PreviewVideo", "PreviewUI3D", "PreviewText", ]
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api/latest/_ui.py", "license": "GNU General Public License v3.0", "lines": 409, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_api/generate_api_stubs.py
#!/usr/bin/env python3 """ Script to generate .pyi stub files for the synchronous API wrappers. This allows generating stubs without running the full ComfyUI application. """ import os import sys import logging import importlib # Add ComfyUI to path so we can import modules sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) from comfy_api.internal.async_to_sync import AsyncToSyncConverter from comfy_api.version_list import supported_versions def generate_stubs_for_module(module_name: str) -> None: """Generate stub files for a specific module that exports ComfyAPI and ComfyAPISync.""" try: # Import the module module = importlib.import_module(module_name) # Check if module has ComfyAPISync (the sync wrapper) if hasattr(module, "ComfyAPISync"): # Module already has a sync class api_class = getattr(module, "ComfyAPI", None) sync_class = getattr(module, "ComfyAPISync") if api_class: # Generate the stub file AsyncToSyncConverter.generate_stub_file(api_class, sync_class) logging.info(f"Generated stub file for {module_name}") else: logging.warning( f"Module {module_name} has ComfyAPISync but no ComfyAPI" ) elif hasattr(module, "ComfyAPI"): # Module only has async API, need to create sync wrapper first from comfy_api.internal.async_to_sync import create_sync_class api_class = getattr(module, "ComfyAPI") sync_class = create_sync_class(api_class) # Generate the stub file AsyncToSyncConverter.generate_stub_file(api_class, sync_class) logging.info(f"Generated stub file for {module_name}") else: logging.warning( f"Module {module_name} does not export ComfyAPI or ComfyAPISync" ) except Exception as e: logging.error(f"Failed to generate stub for {module_name}: {e}") import traceback traceback.print_exc() def main(): """Main function to generate all API stub files.""" logging.basicConfig(level=logging.INFO) logging.info("Starting stub generation...") # Dynamically get module names from supported_versions api_modules = [] for api_class in supported_versions: # Extract module name from the class module_name = api_class.__module__ if module_name not in api_modules: api_modules.append(module_name) logging.info(f"Found {len(api_modules)} API modules: {api_modules}") # Generate stubs for each module for module_name in api_modules: generate_stubs_for_module(module_name) logging.info("Stub generation complete!") if __name__ == "__main__": main()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api/generate_api_stubs.py", "license": "GNU General Public License v3.0", "lines": 65, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_api/internal/api_registry.py
from typing import NamedTuple from comfy_api.internal.singleton import ProxiedSingleton from packaging import version as packaging_version class ComfyAPIBase(ProxiedSingleton): def __init__(self): pass class ComfyAPIWithVersion(NamedTuple): version: str api_class: type[ComfyAPIBase] def parse_version(version_str: str) -> packaging_version.Version: """ Parses a version string into a packaging_version.Version object. Raises ValueError if the version string is invalid. """ if version_str == "latest": return packaging_version.parse("9999999.9999999.9999999") return packaging_version.parse(version_str) registered_versions: list[ComfyAPIWithVersion] = [] def register_versions(versions: list[ComfyAPIWithVersion]): versions.sort(key=lambda x: parse_version(x.version)) global registered_versions registered_versions = versions def get_all_versions() -> list[ComfyAPIWithVersion]: """ Returns a list of all registered ComfyAPI versions. """ return registered_versions
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api/internal/api_registry.py", "license": "GNU General Public License v3.0", "lines": 27, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy_api/internal/singleton.py
from typing import TypeVar class SingletonMetaclass(type): T = TypeVar("T", bound="SingletonMetaclass") _instances = {} def __call__(cls, *args, **kwargs): if cls not in cls._instances: cls._instances[cls] = super(SingletonMetaclass, cls).__call__( *args, **kwargs ) return cls._instances[cls] def inject_instance(cls: type[T], instance: T) -> None: assert cls not in SingletonMetaclass._instances, ( "Cannot inject instance after first instantiation" ) SingletonMetaclass._instances[cls] = instance def get_instance(cls: type[T], *args, **kwargs) -> T: """ Gets the singleton instance of the class, creating it if it doesn't exist. """ if cls not in SingletonMetaclass._instances: SingletonMetaclass._instances[cls] = super( SingletonMetaclass, cls ).__call__(*args, **kwargs) return cls._instances[cls] class ProxiedSingleton(object, metaclass=SingletonMetaclass): def __init__(self): super().__init__()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api/internal/singleton.py", "license": "GNU General Public License v3.0", "lines": 27, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy_api/latest/_input/basic_types.py
import torch from typing import TypedDict, Optional ImageInput = torch.Tensor """ An image in format [B, H, W, C] where B is the batch size, C is the number of channels, """ MaskInput = torch.Tensor """ A mask in format [B, H, W] where B is the batch size """ class AudioInput(TypedDict): """ TypedDict representing audio input. """ waveform: torch.Tensor """ Tensor in the format [B, C, T] where B is the batch size, C is the number of channels, """ sample_rate: int class LatentInput(TypedDict): """ TypedDict representing latent input. """ samples: torch.Tensor """ Tensor in the format [B, C, H, W] where B is the batch size, C is the number of channels, H is the height, and W is the width. """ noise_mask: Optional[MaskInput] """ Optional noise mask tensor in the same format as samples. """ batch_index: Optional[list[int]]
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api/latest/_input/basic_types.py", "license": "GNU General Public License v3.0", "lines": 33, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
documentation
Comfy-Org/ComfyUI:comfy_api/latest/_input/video_types.py
from __future__ import annotations from abc import ABC, abstractmethod from fractions import Fraction from typing import Optional, Union, IO import io import av from .._util import VideoContainer, VideoCodec, VideoComponents class VideoInput(ABC): """ Abstract base class for video input types. """ @abstractmethod def get_components(self) -> VideoComponents: """ Abstract method to get the video components (images, audio, and frame rate). Returns: VideoComponents containing images, audio, and frame rate """ pass @abstractmethod def save_to( self, path: Union[str, IO[bytes]], format: VideoContainer = VideoContainer.AUTO, codec: VideoCodec = VideoCodec.AUTO, metadata: Optional[dict] = None ): """ Abstract method to save the video input to a file. """ pass @abstractmethod def as_trimmed( self, start_time: float | None = None, duration: float | None = None, strict_duration: bool = False, ) -> VideoInput | None: """ Create a new VideoInput which is trimmed to have the corresponding start_time and duration Returns: A new VideoInput, or None if the result would have negative duration """ pass def get_stream_source(self) -> Union[str, io.BytesIO]: """ Get a streamable source for the video. This allows processing without loading the entire video into memory. Returns: Either a file path (str) or a BytesIO object that can be opened with av. Default implementation creates a BytesIO buffer, but subclasses should override this for better performance when possible. """ buffer = io.BytesIO() self.save_to(buffer) buffer.seek(0) return buffer # Provide a default implementation, but subclasses can provide optimized versions # if possible. def get_dimensions(self) -> tuple[int, int]: """ Returns the dimensions of the video input. Returns: Tuple of (width, height) """ components = self.get_components() return components.images.shape[2], components.images.shape[1] def get_duration(self) -> float: """ Returns the duration of the video in seconds. Returns: Duration in seconds """ components = self.get_components() frame_count = components.images.shape[0] return float(frame_count / components.frame_rate) def get_frame_count(self) -> int: """ Returns the number of frames in the video. Default implementation uses :meth:`get_components`, which may require loading all frames into memory. File-based implementations should override this method and use container/stream metadata instead. Returns: Total number of frames as an integer. """ return int(self.get_components().images.shape[0]) def get_frame_rate(self) -> Fraction: """ Returns the frame rate of the video. Default implementation materializes the video into memory via `get_components()`. Subclasses that can inspect the underlying container (e.g. `VideoFromFile`) should override this with a more efficient implementation. Returns: Frame rate as a Fraction. """ return self.get_components().frame_rate def get_container_format(self) -> str: """ Returns the container format of the video (e.g., 'mp4', 'mov', 'avi'). Returns: Container format as string """ # Default implementation - subclasses should override for better performance source = self.get_stream_source() with av.open(source, mode="r") as container: return container.format.name
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api/latest/_input/video_types.py", "license": "GNU General Public License v3.0", "lines": 107, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
documentation
Comfy-Org/ComfyUI:comfy_api/latest/_util/video_types.py
from __future__ import annotations from dataclasses import dataclass from enum import Enum from fractions import Fraction from typing import Optional from .._input import ImageInput, AudioInput class VideoCodec(str, Enum): AUTO = "auto" H264 = "h264" @classmethod def as_input(cls) -> list[str]: """ Returns a list of codec names that can be used as node input. """ return [member.value for member in cls] class VideoContainer(str, Enum): AUTO = "auto" MP4 = "mp4" @classmethod def as_input(cls) -> list[str]: """ Returns a list of container names that can be used as node input. """ return [member.value for member in cls] @classmethod def get_extension(cls, value) -> str: """ Returns the file extension for the container. """ if isinstance(value, str): value = cls(value) if value == VideoContainer.MP4 or value == VideoContainer.AUTO: return "mp4" return "" @dataclass class VideoComponents: """ Dataclass representing the components of a video. """ images: ImageInput frame_rate: Fraction audio: Optional[AudioInput] = None metadata: Optional[dict] = None
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api/latest/_util/video_types.py", "license": "GNU General Public License v3.0", "lines": 43, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy_api/util.py
# This file only exists for backwards compatibility. from comfy_api.latest._util import VideoCodec, VideoContainer, VideoComponents __all__ = [ "VideoCodec", "VideoContainer", "VideoComponents", ]
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api/util.py", "license": "GNU General Public License v3.0", "lines": 7, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy_api/version_list.py
from comfy_api.latest import ComfyAPI_latest from comfy_api.v0_0_2 import ComfyAPIAdapter_v0_0_2 from comfy_api.v0_0_1 import ComfyAPIAdapter_v0_0_1 from comfy_api.internal import ComfyAPIBase supported_versions: list[type[ComfyAPIBase]] = [ ComfyAPI_latest, ComfyAPIAdapter_v0_0_2, ComfyAPIAdapter_v0_0_1, ]
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api/version_list.py", "license": "GNU General Public License v3.0", "lines": 9, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy/ldm/wan/vae2_2.py
# original version: https://github.com/Wan-Video/Wan2.2/blob/main/wan/modules/vae2_2.py # Copyright 2024-2025 The Alibaba Wan Team Authors. All rights reserved. import torch import torch.nn as nn import torch.nn.functional as F from einops import rearrange from .vae import AttentionBlock, CausalConv3d, RMS_norm import comfy.ops ops = comfy.ops.disable_weight_init CACHE_T = 2 class Resample(nn.Module): def __init__(self, dim, mode): assert mode in ( "none", "upsample2d", "upsample3d", "downsample2d", "downsample3d", ) super().__init__() self.dim = dim self.mode = mode # layers if mode == "upsample2d": self.resample = nn.Sequential( nn.Upsample(scale_factor=(2.0, 2.0), mode="nearest-exact"), ops.Conv2d(dim, dim, 3, padding=1), ) elif mode == "upsample3d": self.resample = nn.Sequential( nn.Upsample(scale_factor=(2.0, 2.0), mode="nearest-exact"), ops.Conv2d(dim, dim, 3, padding=1), # ops.Conv2d(dim, dim//2, 3, padding=1) ) self.time_conv = CausalConv3d( dim, dim * 2, (3, 1, 1), padding=(1, 0, 0)) elif mode == "downsample2d": self.resample = nn.Sequential( nn.ZeroPad2d((0, 1, 0, 1)), ops.Conv2d(dim, dim, 3, stride=(2, 2))) elif mode == "downsample3d": self.resample = nn.Sequential( nn.ZeroPad2d((0, 1, 0, 1)), ops.Conv2d(dim, dim, 3, stride=(2, 2))) self.time_conv = CausalConv3d( dim, dim, (3, 1, 1), stride=(2, 1, 1), padding=(0, 0, 0)) else: self.resample = nn.Identity() def forward(self, x, feat_cache=None, feat_idx=[0]): b, c, t, h, w = x.size() if self.mode == "upsample3d": if feat_cache is not None: idx = feat_idx[0] if feat_cache[idx] is None: feat_cache[idx] = "Rep" feat_idx[0] += 1 else: cache_x = x[:, :, -CACHE_T:, :, :].clone() if (cache_x.shape[2] < 2 and feat_cache[idx] is not None and feat_cache[idx] != "Rep"): # cache last frame of last two chunk cache_x = torch.cat( [ feat_cache[idx][:, :, -1, :, :].unsqueeze(2).to( cache_x.device), cache_x, ], dim=2, ) if (cache_x.shape[2] < 2 and feat_cache[idx] is not None and feat_cache[idx] == "Rep"): cache_x = torch.cat( [ torch.zeros_like(cache_x).to(cache_x.device), cache_x ], dim=2, ) if feat_cache[idx] == "Rep": x = self.time_conv(x) else: x = self.time_conv(x, feat_cache[idx]) feat_cache[idx] = cache_x feat_idx[0] += 1 x = x.reshape(b, 2, c, t, h, w) x = torch.stack((x[:, 0, :, :, :, :], x[:, 1, :, :, :, :]), 3) x = x.reshape(b, c, t * 2, h, w) t = x.shape[2] x = rearrange(x, "b c t h w -> (b t) c h w") x = self.resample(x) x = rearrange(x, "(b t) c h w -> b c t h w", t=t) if self.mode == "downsample3d": if feat_cache is not None: idx = feat_idx[0] if feat_cache[idx] is None: feat_cache[idx] = x.clone() feat_idx[0] += 1 else: cache_x = x[:, :, -1:, :, :].clone() x = self.time_conv( torch.cat([feat_cache[idx][:, :, -1:, :, :], x], 2)) feat_cache[idx] = cache_x feat_idx[0] += 1 return x class ResidualBlock(nn.Module): def __init__(self, in_dim, out_dim, dropout=0.0): super().__init__() self.in_dim = in_dim self.out_dim = out_dim # layers self.residual = nn.Sequential( RMS_norm(in_dim, images=False), nn.SiLU(), CausalConv3d(in_dim, out_dim, 3, padding=1), RMS_norm(out_dim, images=False), nn.SiLU(), nn.Dropout(dropout), CausalConv3d(out_dim, out_dim, 3, padding=1), ) self.shortcut = ( CausalConv3d(in_dim, out_dim, 1) if in_dim != out_dim else nn.Identity()) def forward(self, x, feat_cache=None, feat_idx=[0]): old_x = x for layer in self.residual: if isinstance(layer, CausalConv3d) and feat_cache is not None: idx = feat_idx[0] cache_x = x[:, :, -CACHE_T:, :, :].clone() if cache_x.shape[2] < 2 and feat_cache[idx] is not None: # cache last frame of last two chunk cache_x = torch.cat( [ feat_cache[idx][:, :, -1, :, :].unsqueeze(2).to( cache_x.device), cache_x, ], dim=2, ) x = layer(x, cache_list=feat_cache, cache_idx=idx) feat_cache[idx] = cache_x feat_idx[0] += 1 else: x = layer(x) return x + self.shortcut(old_x) def patchify(x, patch_size): if patch_size == 1: return x if x.dim() == 4: x = rearrange( x, "b c (h q) (w r) -> b (c r q) h w", q=patch_size, r=patch_size) elif x.dim() == 5: x = rearrange( x, "b c f (h q) (w r) -> b (c r q) f h w", q=patch_size, r=patch_size, ) else: raise ValueError(f"Invalid input shape: {x.shape}") return x def unpatchify(x, patch_size): if patch_size == 1: return x if x.dim() == 4: x = rearrange( x, "b (c r q) h w -> b c (h q) (w r)", q=patch_size, r=patch_size) elif x.dim() == 5: x = rearrange( x, "b (c r q) f h w -> b c f (h q) (w r)", q=patch_size, r=patch_size, ) return x class AvgDown3D(nn.Module): def __init__( self, in_channels, out_channels, factor_t, factor_s=1, ): super().__init__() self.in_channels = in_channels self.out_channels = out_channels self.factor_t = factor_t self.factor_s = factor_s self.factor = self.factor_t * self.factor_s * self.factor_s assert in_channels * self.factor % out_channels == 0 self.group_size = in_channels * self.factor // out_channels def forward(self, x: torch.Tensor) -> torch.Tensor: pad_t = (self.factor_t - x.shape[2] % self.factor_t) % self.factor_t pad = (0, 0, 0, 0, pad_t, 0) x = F.pad(x, pad) B, C, T, H, W = x.shape x = x.view( B, C, T // self.factor_t, self.factor_t, H // self.factor_s, self.factor_s, W // self.factor_s, self.factor_s, ) x = x.permute(0, 1, 3, 5, 7, 2, 4, 6).contiguous() x = x.view( B, C * self.factor, T // self.factor_t, H // self.factor_s, W // self.factor_s, ) x = x.view( B, self.out_channels, self.group_size, T // self.factor_t, H // self.factor_s, W // self.factor_s, ) x = x.mean(dim=2) return x class DupUp3D(nn.Module): def __init__( self, in_channels: int, out_channels: int, factor_t, factor_s=1, ): super().__init__() self.in_channels = in_channels self.out_channels = out_channels self.factor_t = factor_t self.factor_s = factor_s self.factor = self.factor_t * self.factor_s * self.factor_s assert out_channels * self.factor % in_channels == 0 self.repeats = out_channels * self.factor // in_channels def forward(self, x: torch.Tensor, first_chunk=False) -> torch.Tensor: x = x.repeat_interleave(self.repeats, dim=1) x = x.view( x.size(0), self.out_channels, self.factor_t, self.factor_s, self.factor_s, x.size(2), x.size(3), x.size(4), ) x = x.permute(0, 1, 5, 2, 6, 3, 7, 4).contiguous() x = x.view( x.size(0), self.out_channels, x.size(2) * self.factor_t, x.size(4) * self.factor_s, x.size(6) * self.factor_s, ) if first_chunk: x = x[:, :, self.factor_t - 1:, :, :] return x class Down_ResidualBlock(nn.Module): def __init__(self, in_dim, out_dim, dropout, mult, temperal_downsample=False, down_flag=False): super().__init__() # Shortcut path with downsample self.avg_shortcut = AvgDown3D( in_dim, out_dim, factor_t=2 if temperal_downsample else 1, factor_s=2 if down_flag else 1, ) # Main path with residual blocks and downsample downsamples = [] for _ in range(mult): downsamples.append(ResidualBlock(in_dim, out_dim, dropout)) in_dim = out_dim # Add the final downsample block if down_flag: mode = "downsample3d" if temperal_downsample else "downsample2d" downsamples.append(Resample(out_dim, mode=mode)) self.downsamples = nn.Sequential(*downsamples) def forward(self, x, feat_cache=None, feat_idx=[0]): x_copy = x for module in self.downsamples: x = module(x, feat_cache, feat_idx) return x + self.avg_shortcut(x_copy) class Up_ResidualBlock(nn.Module): def __init__(self, in_dim, out_dim, dropout, mult, temperal_upsample=False, up_flag=False): super().__init__() # Shortcut path with upsample if up_flag: self.avg_shortcut = DupUp3D( in_dim, out_dim, factor_t=2 if temperal_upsample else 1, factor_s=2 if up_flag else 1, ) else: self.avg_shortcut = None # Main path with residual blocks and upsample upsamples = [] for _ in range(mult): upsamples.append(ResidualBlock(in_dim, out_dim, dropout)) in_dim = out_dim # Add the final upsample block if up_flag: mode = "upsample3d" if temperal_upsample else "upsample2d" upsamples.append(Resample(out_dim, mode=mode)) self.upsamples = nn.Sequential(*upsamples) def forward(self, x, feat_cache=None, feat_idx=[0], first_chunk=False): x_main = x for module in self.upsamples: x_main = module(x_main, feat_cache, feat_idx) if self.avg_shortcut is not None: x_shortcut = self.avg_shortcut(x, first_chunk) return x_main + x_shortcut else: return x_main class Encoder3d(nn.Module): def __init__( self, dim=128, z_dim=4, dim_mult=[1, 2, 4, 4], num_res_blocks=2, attn_scales=[], temperal_downsample=[True, True, False], dropout=0.0, ): super().__init__() self.dim = dim self.z_dim = z_dim self.dim_mult = dim_mult self.num_res_blocks = num_res_blocks self.attn_scales = attn_scales self.temperal_downsample = temperal_downsample # dimensions dims = [dim * u for u in [1] + dim_mult] scale = 1.0 # init block self.conv1 = CausalConv3d(12, dims[0], 3, padding=1) # downsample blocks downsamples = [] for i, (in_dim, out_dim) in enumerate(zip(dims[:-1], dims[1:])): t_down_flag = ( temperal_downsample[i] if i < len(temperal_downsample) else False) downsamples.append( Down_ResidualBlock( in_dim=in_dim, out_dim=out_dim, dropout=dropout, mult=num_res_blocks, temperal_downsample=t_down_flag, down_flag=i != len(dim_mult) - 1, )) scale /= 2.0 self.downsamples = nn.Sequential(*downsamples) # middle blocks self.middle = nn.Sequential( ResidualBlock(out_dim, out_dim, dropout), AttentionBlock(out_dim), ResidualBlock(out_dim, out_dim, dropout), ) # # output blocks self.head = nn.Sequential( RMS_norm(out_dim, images=False), nn.SiLU(), CausalConv3d(out_dim, z_dim, 3, padding=1), ) def forward(self, x, feat_cache=None, feat_idx=[0]): if feat_cache is not None: idx = feat_idx[0] cache_x = x[:, :, -CACHE_T:, :, :].clone() if cache_x.shape[2] < 2 and feat_cache[idx] is not None: cache_x = torch.cat( [ feat_cache[idx][:, :, -1, :, :].unsqueeze(2).to( cache_x.device), cache_x, ], dim=2, ) x = self.conv1(x, feat_cache[idx]) feat_cache[idx] = cache_x feat_idx[0] += 1 else: x = self.conv1(x) ## downsamples for layer in self.downsamples: if feat_cache is not None: x = layer(x, feat_cache, feat_idx) else: x = layer(x) ## middle for layer in self.middle: if isinstance(layer, ResidualBlock) and feat_cache is not None: x = layer(x, feat_cache, feat_idx) else: x = layer(x) ## head for layer in self.head: if isinstance(layer, CausalConv3d) and feat_cache is not None: idx = feat_idx[0] cache_x = x[:, :, -CACHE_T:, :, :].clone() if cache_x.shape[2] < 2 and feat_cache[idx] is not None: cache_x = torch.cat( [ feat_cache[idx][:, :, -1, :, :].unsqueeze(2).to( cache_x.device), cache_x, ], dim=2, ) x = layer(x, feat_cache[idx]) feat_cache[idx] = cache_x feat_idx[0] += 1 else: x = layer(x) return x class Decoder3d(nn.Module): def __init__( self, dim=128, z_dim=4, dim_mult=[1, 2, 4, 4], num_res_blocks=2, attn_scales=[], temperal_upsample=[False, True, True], dropout=0.0, ): super().__init__() self.dim = dim self.z_dim = z_dim self.dim_mult = dim_mult self.num_res_blocks = num_res_blocks self.attn_scales = attn_scales self.temperal_upsample = temperal_upsample # dimensions dims = [dim * u for u in [dim_mult[-1]] + dim_mult[::-1]] # init block self.conv1 = CausalConv3d(z_dim, dims[0], 3, padding=1) # middle blocks self.middle = nn.Sequential( ResidualBlock(dims[0], dims[0], dropout), AttentionBlock(dims[0]), ResidualBlock(dims[0], dims[0], dropout), ) # upsample blocks upsamples = [] for i, (in_dim, out_dim) in enumerate(zip(dims[:-1], dims[1:])): t_up_flag = temperal_upsample[i] if i < len( temperal_upsample) else False upsamples.append( Up_ResidualBlock( in_dim=in_dim, out_dim=out_dim, dropout=dropout, mult=num_res_blocks + 1, temperal_upsample=t_up_flag, up_flag=i != len(dim_mult) - 1, )) self.upsamples = nn.Sequential(*upsamples) # output blocks self.head = nn.Sequential( RMS_norm(out_dim, images=False), nn.SiLU(), CausalConv3d(out_dim, 12, 3, padding=1), ) def forward(self, x, feat_cache=None, feat_idx=[0], first_chunk=False): if feat_cache is not None: idx = feat_idx[0] cache_x = x[:, :, -CACHE_T:, :, :].clone() if cache_x.shape[2] < 2 and feat_cache[idx] is not None: cache_x = torch.cat( [ feat_cache[idx][:, :, -1, :, :].unsqueeze(2).to( cache_x.device), cache_x, ], dim=2, ) x = self.conv1(x, feat_cache[idx]) feat_cache[idx] = cache_x feat_idx[0] += 1 else: x = self.conv1(x) for layer in self.middle: if isinstance(layer, ResidualBlock) and feat_cache is not None: x = layer(x, feat_cache, feat_idx) else: x = layer(x) ## upsamples for layer in self.upsamples: if feat_cache is not None: x = layer(x, feat_cache, feat_idx, first_chunk) else: x = layer(x) ## head for layer in self.head: if isinstance(layer, CausalConv3d) and feat_cache is not None: idx = feat_idx[0] cache_x = x[:, :, -CACHE_T:, :, :].clone() if cache_x.shape[2] < 2 and feat_cache[idx] is not None: cache_x = torch.cat( [ feat_cache[idx][:, :, -1, :, :].unsqueeze(2).to( cache_x.device), cache_x, ], dim=2, ) x = layer(x, feat_cache[idx]) feat_cache[idx] = cache_x feat_idx[0] += 1 else: x = layer(x) return x def count_conv3d(model): count = 0 for m in model.modules(): if isinstance(m, CausalConv3d): count += 1 return count class WanVAE(nn.Module): def __init__( self, dim=160, dec_dim=256, z_dim=16, dim_mult=[1, 2, 4, 4], num_res_blocks=2, attn_scales=[], temperal_downsample=[True, True, False], dropout=0.0, ): super().__init__() self.dim = dim self.z_dim = z_dim self.dim_mult = dim_mult self.num_res_blocks = num_res_blocks self.attn_scales = attn_scales self.temperal_downsample = temperal_downsample self.temperal_upsample = temperal_downsample[::-1] # modules self.encoder = Encoder3d( dim, z_dim * 2, dim_mult, num_res_blocks, attn_scales, self.temperal_downsample, dropout, ) self.conv1 = CausalConv3d(z_dim * 2, z_dim * 2, 1) self.conv2 = CausalConv3d(z_dim, z_dim, 1) self.decoder = Decoder3d( dec_dim, z_dim, dim_mult, num_res_blocks, attn_scales, self.temperal_upsample, dropout, ) def encode(self, x): conv_idx = [0] feat_map = [None] * count_conv3d(self.encoder) x = patchify(x, patch_size=2) t = x.shape[2] iter_ = 1 + (t - 1) // 4 for i in range(iter_): conv_idx = [0] if i == 0: out = self.encoder( x[:, :, :1, :, :], feat_cache=feat_map, feat_idx=conv_idx, ) else: out_ = self.encoder( x[:, :, 1 + 4 * (i - 1):1 + 4 * i, :, :], feat_cache=feat_map, feat_idx=conv_idx, ) out = torch.cat([out, out_], 2) mu, log_var = self.conv1(out).chunk(2, dim=1) return mu def decode(self, z): conv_idx = [0] feat_map = [None] * count_conv3d(self.decoder) iter_ = z.shape[2] x = self.conv2(z) for i in range(iter_): conv_idx = [0] if i == 0: out = self.decoder( x[:, :, i:i + 1, :, :], feat_cache=feat_map, feat_idx=conv_idx, first_chunk=True, ) else: out_ = self.decoder( x[:, :, i:i + 1, :, :], feat_cache=feat_map, feat_idx=conv_idx, ) out = torch.cat([out, out_], 2) out = unpatchify(out, patch_size=2) return out def reparameterize(self, mu, log_var): std = torch.exp(0.5 * log_var) eps = torch.randn_like(std) return eps * std + mu def sample(self, imgs, deterministic=False): mu, log_var = self.encode(imgs) if deterministic: return mu std = torch.exp(0.5 * log_var.clamp(-30.0, 20.0)) return mu + std * torch.randn_like(std)
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/ldm/wan/vae2_2.py", "license": "GNU General Public License v3.0", "lines": 635, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_execution/progress.py
from __future__ import annotations from typing import TypedDict, Dict, Optional, Tuple from typing_extensions import override from PIL import Image from enum import Enum from abc import ABC from tqdm import tqdm from typing import TYPE_CHECKING if TYPE_CHECKING: from comfy_execution.graph import DynamicPrompt from protocol import BinaryEventTypes from comfy_api import feature_flags PreviewImageTuple = Tuple[str, Image.Image, Optional[int]] class NodeState(Enum): Pending = "pending" Running = "running" Finished = "finished" Error = "error" class NodeProgressState(TypedDict): """ A class to represent the state of a node's progress. """ state: NodeState value: float max: float class ProgressHandler(ABC): """ Abstract base class for progress handlers. Progress handlers receive progress updates and display them in various ways. """ def __init__(self, name: str): self.name = name self.enabled = True def set_registry(self, registry: "ProgressRegistry"): pass def start_handler(self, node_id: str, state: NodeProgressState, prompt_id: str): """Called when a node starts processing""" pass def update_handler( self, node_id: str, value: float, max_value: float, state: NodeProgressState, prompt_id: str, image: PreviewImageTuple | None = None, ): """Called when a node's progress is updated""" pass def finish_handler(self, node_id: str, state: NodeProgressState, prompt_id: str): """Called when a node finishes processing""" pass def reset(self): """Called when the progress registry is reset""" pass def enable(self): """Enable this handler""" self.enabled = True def disable(self): """Disable this handler""" self.enabled = False class CLIProgressHandler(ProgressHandler): """ Handler that displays progress using tqdm progress bars in the CLI. """ def __init__(self): super().__init__("cli") self.progress_bars: Dict[str, tqdm] = {} @override def start_handler(self, node_id: str, state: NodeProgressState, prompt_id: str): # Create a new tqdm progress bar if node_id not in self.progress_bars: self.progress_bars[node_id] = tqdm( total=state["max"], desc=f"Node {node_id}", unit="steps", leave=True, position=len(self.progress_bars), ) @override def update_handler( self, node_id: str, value: float, max_value: float, state: NodeProgressState, prompt_id: str, image: PreviewImageTuple | None = None, ): # Handle case where start_handler wasn't called if node_id not in self.progress_bars: self.progress_bars[node_id] = tqdm( total=max_value, desc=f"Node {node_id}", unit="steps", leave=True, position=len(self.progress_bars), ) self.progress_bars[node_id].update(value) else: # Update existing progress bar if max_value != self.progress_bars[node_id].total: self.progress_bars[node_id].total = max_value # Calculate the update amount (difference from current position) current_position = self.progress_bars[node_id].n update_amount = value - current_position if update_amount > 0: self.progress_bars[node_id].update(update_amount) @override def finish_handler(self, node_id: str, state: NodeProgressState, prompt_id: str): # Complete and close the progress bar if it exists if node_id in self.progress_bars: # Ensure the bar shows 100% completion remaining = state["max"] - self.progress_bars[node_id].n if remaining > 0: self.progress_bars[node_id].update(remaining) self.progress_bars[node_id].close() del self.progress_bars[node_id] @override def reset(self): # Close all progress bars for bar in self.progress_bars.values(): bar.close() self.progress_bars.clear() class WebUIProgressHandler(ProgressHandler): """ Handler that sends progress updates to the WebUI via WebSockets. """ def __init__(self, server_instance): super().__init__("webui") self.server_instance = server_instance def set_registry(self, registry: "ProgressRegistry"): self.registry = registry def _send_progress_state(self, prompt_id: str, nodes: Dict[str, NodeProgressState]): """Send the current progress state to the client""" if self.server_instance is None: return # Only send info for non-pending nodes active_nodes = { node_id: { "value": state["value"], "max": state["max"], "state": state["state"].value, "node_id": node_id, "prompt_id": prompt_id, "display_node_id": self.registry.dynprompt.get_display_node_id(node_id), "parent_node_id": self.registry.dynprompt.get_parent_node_id(node_id), "real_node_id": self.registry.dynprompt.get_real_node_id(node_id), } for node_id, state in nodes.items() if state["state"] != NodeState.Pending } # Send a combined progress_state message with all node states # Include client_id to ensure message is only sent to the initiating client self.server_instance.send_sync( "progress_state", {"prompt_id": prompt_id, "nodes": active_nodes}, self.server_instance.client_id ) @override def start_handler(self, node_id: str, state: NodeProgressState, prompt_id: str): # Send progress state of all nodes if self.registry: self._send_progress_state(prompt_id, self.registry.nodes) @override def update_handler( self, node_id: str, value: float, max_value: float, state: NodeProgressState, prompt_id: str, image: PreviewImageTuple | None = None, ): # Send progress state of all nodes if self.registry: self._send_progress_state(prompt_id, self.registry.nodes) if image: # Only send new format if client supports it if feature_flags.supports_feature( self.server_instance.sockets_metadata, self.server_instance.client_id, "supports_preview_metadata", ): metadata = { "node_id": node_id, "prompt_id": prompt_id, "display_node_id": self.registry.dynprompt.get_display_node_id( node_id ), "parent_node_id": self.registry.dynprompt.get_parent_node_id( node_id ), "real_node_id": self.registry.dynprompt.get_real_node_id(node_id), } self.server_instance.send_sync( BinaryEventTypes.PREVIEW_IMAGE_WITH_METADATA, (image, metadata), self.server_instance.client_id, ) @override def finish_handler(self, node_id: str, state: NodeProgressState, prompt_id: str): # Send progress state of all nodes if self.registry: self._send_progress_state(prompt_id, self.registry.nodes) class ProgressRegistry: """ Registry that maintains node progress state and notifies registered handlers. """ def __init__(self, prompt_id: str, dynprompt: "DynamicPrompt"): self.prompt_id = prompt_id self.dynprompt = dynprompt self.nodes: Dict[str, NodeProgressState] = {} self.handlers: Dict[str, ProgressHandler] = {} def register_handler(self, handler: ProgressHandler) -> None: """Register a progress handler""" self.handlers[handler.name] = handler def unregister_handler(self, handler_name: str) -> None: """Unregister a progress handler""" if handler_name in self.handlers: # Allow handler to clean up resources self.handlers[handler_name].reset() del self.handlers[handler_name] def enable_handler(self, handler_name: str) -> None: """Enable a progress handler""" if handler_name in self.handlers: self.handlers[handler_name].enable() def disable_handler(self, handler_name: str) -> None: """Disable a progress handler""" if handler_name in self.handlers: self.handlers[handler_name].disable() def ensure_entry(self, node_id: str) -> NodeProgressState: """Ensure a node entry exists""" if node_id not in self.nodes: self.nodes[node_id] = NodeProgressState( state=NodeState.Pending, value=0, max=1 ) return self.nodes[node_id] def start_progress(self, node_id: str) -> None: """Start progress tracking for a node""" entry = self.ensure_entry(node_id) entry["state"] = NodeState.Running entry["value"] = 0.0 entry["max"] = 1.0 # Notify all enabled handlers for handler in self.handlers.values(): if handler.enabled: handler.start_handler(node_id, entry, self.prompt_id) def update_progress( self, node_id: str, value: float, max_value: float, image: PreviewImageTuple | None = None ) -> None: """Update progress for a node""" entry = self.ensure_entry(node_id) entry["state"] = NodeState.Running entry["value"] = value entry["max"] = max_value # Notify all enabled handlers for handler in self.handlers.values(): if handler.enabled: handler.update_handler( node_id, value, max_value, entry, self.prompt_id, image ) def finish_progress(self, node_id: str) -> None: """Finish progress tracking for a node""" entry = self.ensure_entry(node_id) entry["state"] = NodeState.Finished entry["value"] = entry["max"] # Notify all enabled handlers for handler in self.handlers.values(): if handler.enabled: handler.finish_handler(node_id, entry, self.prompt_id) def reset_handlers(self) -> None: """Reset all handlers""" for handler in self.handlers.values(): handler.reset() # Global registry instance global_progress_registry: ProgressRegistry | None = None def reset_progress_state(prompt_id: str, dynprompt: "DynamicPrompt") -> None: global global_progress_registry # Reset existing handlers if registry exists if global_progress_registry is not None: global_progress_registry.reset_handlers() # Create new registry global_progress_registry = ProgressRegistry(prompt_id, dynprompt) def add_progress_handler(handler: ProgressHandler) -> None: registry = get_progress_state() handler.set_registry(registry) registry.register_handler(handler) def get_progress_state() -> ProgressRegistry: global global_progress_registry if global_progress_registry is None: from comfy_execution.graph import DynamicPrompt global_progress_registry = ProgressRegistry( prompt_id="", dynprompt=DynamicPrompt({}) ) return global_progress_registry
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_execution/progress.py", "license": "GNU General Public License v3.0", "lines": 294, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_execution/utils.py
import contextvars from typing import Optional, NamedTuple class ExecutionContext(NamedTuple): """ Context information about the currently executing node. Attributes: node_id: The ID of the currently executing node list_index: The index in a list being processed (for operations on batches/lists) """ prompt_id: str node_id: str list_index: Optional[int] current_executing_context: contextvars.ContextVar[Optional[ExecutionContext]] = contextvars.ContextVar("current_executing_context", default=None) def get_executing_context() -> Optional[ExecutionContext]: return current_executing_context.get(None) class CurrentNodeContext: """ Context manager for setting the current executing node context. Sets the current_executing_context on enter and resets it on exit. Example: with CurrentNodeContext(node_id="123", list_index=0): # Code that should run with the current node context set process_image() """ def __init__(self, prompt_id: str, node_id: str, list_index: Optional[int] = None): self.context = ExecutionContext( prompt_id= prompt_id, node_id= node_id, list_index= list_index ) self.token = None def __enter__(self): self.token = current_executing_context.set(self.context) return self def __exit__(self, exc_type, exc_val, exc_tb): if self.token is not None: current_executing_context.reset(self.token)
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_execution/utils.py", "license": "GNU General Public License v3.0", "lines": 37, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:protocol.py
class BinaryEventTypes: PREVIEW_IMAGE = 1 UNENCODED_PREVIEW_IMAGE = 2 TEXT = 3 PREVIEW_IMAGE_WITH_METADATA = 4
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "protocol.py", "license": "GNU General Public License v3.0", "lines": 5, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:tests-unit/feature_flags_test.py
"""Tests for feature flags functionality.""" from comfy_api.feature_flags import ( get_connection_feature, supports_feature, get_server_features, SERVER_FEATURE_FLAGS, ) class TestFeatureFlags: """Test suite for feature flags functions.""" def test_get_server_features_returns_copy(self): """Test that get_server_features returns a copy of the server flags.""" features = get_server_features() # Verify it's a copy by modifying it features["test_flag"] = True # Original should be unchanged assert "test_flag" not in SERVER_FEATURE_FLAGS def test_get_server_features_contains_expected_flags(self): """Test that server features contain expected flags.""" features = get_server_features() assert "supports_preview_metadata" in features assert features["supports_preview_metadata"] is True assert "max_upload_size" in features assert isinstance(features["max_upload_size"], (int, float)) def test_get_connection_feature_with_missing_sid(self): """Test getting feature for non-existent session ID.""" sockets_metadata = {} result = get_connection_feature(sockets_metadata, "missing_sid", "some_feature") assert result is False # Default value def test_get_connection_feature_with_custom_default(self): """Test getting feature with custom default value.""" sockets_metadata = {} result = get_connection_feature( sockets_metadata, "missing_sid", "some_feature", default="custom_default" ) assert result == "custom_default" def test_get_connection_feature_with_feature_flags(self): """Test getting feature from connection with feature flags.""" sockets_metadata = { "sid1": { "feature_flags": { "supports_preview_metadata": True, "custom_feature": "value", }, } } result = get_connection_feature(sockets_metadata, "sid1", "supports_preview_metadata") assert result is True result = get_connection_feature(sockets_metadata, "sid1", "custom_feature") assert result == "value" def test_get_connection_feature_missing_feature(self): """Test getting non-existent feature from connection.""" sockets_metadata = { "sid1": {"feature_flags": {"existing_feature": True}} } result = get_connection_feature(sockets_metadata, "sid1", "missing_feature") assert result is False def test_supports_feature_returns_boolean(self): """Test that supports_feature always returns boolean.""" sockets_metadata = { "sid1": { "feature_flags": { "bool_feature": True, "string_feature": "value", "none_feature": None, }, } } # True boolean feature assert supports_feature(sockets_metadata, "sid1", "bool_feature") is True # Non-boolean values should return False assert supports_feature(sockets_metadata, "sid1", "string_feature") is False assert supports_feature(sockets_metadata, "sid1", "none_feature") is False assert supports_feature(sockets_metadata, "sid1", "missing_feature") is False def test_supports_feature_with_missing_connection(self): """Test supports_feature with missing connection.""" sockets_metadata = {} assert supports_feature(sockets_metadata, "missing_sid", "any_feature") is False def test_empty_feature_flags_dict(self): """Test connection with empty feature flags dictionary.""" sockets_metadata = {"sid1": {"feature_flags": {}}} result = get_connection_feature(sockets_metadata, "sid1", "any_feature") assert result is False assert supports_feature(sockets_metadata, "sid1", "any_feature") is False
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "tests-unit/feature_flags_test.py", "license": "GNU General Public License v3.0", "lines": 83, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
Comfy-Org/ComfyUI:tests-unit/websocket_feature_flags_test.py
"""Simplified tests for WebSocket feature flags functionality.""" from comfy_api import feature_flags class TestWebSocketFeatureFlags: """Test suite for WebSocket feature flags integration.""" def test_server_feature_flags_response(self): """Test server feature flags are properly formatted.""" features = feature_flags.get_server_features() # Check expected server features assert "supports_preview_metadata" in features assert features["supports_preview_metadata"] is True assert "max_upload_size" in features assert isinstance(features["max_upload_size"], (int, float)) def test_progress_py_checks_feature_flags(self): """Test that progress.py checks feature flags before sending metadata.""" # This simulates the check in progress.py client_id = "test_client" sockets_metadata = {"test_client": {"feature_flags": {}}} # The actual check would be in progress.py supports_metadata = feature_flags.supports_feature( sockets_metadata, client_id, "supports_preview_metadata" ) assert supports_metadata is False def test_multiple_clients_different_features(self): """Test handling multiple clients with different feature support.""" sockets_metadata = { "modern_client": { "feature_flags": {"supports_preview_metadata": True} }, "legacy_client": { "feature_flags": {} } } # Check modern client assert feature_flags.supports_feature( sockets_metadata, "modern_client", "supports_preview_metadata" ) is True # Check legacy client assert feature_flags.supports_feature( sockets_metadata, "legacy_client", "supports_preview_metadata" ) is False def test_feature_negotiation_message_format(self): """Test the format of feature negotiation messages.""" # Client message format client_message = { "type": "feature_flags", "data": { "supports_preview_metadata": True, "api_version": "1.0.0" } } # Verify structure assert client_message["type"] == "feature_flags" assert "supports_preview_metadata" in client_message["data"] # Server response format (what would be sent) server_features = feature_flags.get_server_features() server_message = { "type": "feature_flags", "data": server_features } # Verify structure assert server_message["type"] == "feature_flags" assert "supports_preview_metadata" in server_message["data"] assert server_message["data"]["supports_preview_metadata"] is True
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "tests-unit/websocket_feature_flags_test.py", "license": "GNU General Public License v3.0", "lines": 63, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
Comfy-Org/ComfyUI:comfy/k_diffusion/sa_solver.py
# SA-Solver: Stochastic Adams Solver (NeurIPS 2023, arXiv:2309.05019) # Conference: https://proceedings.neurips.cc/paper_files/paper/2023/file/f4a6806490d31216a3ba667eb240c897-Paper-Conference.pdf # Codebase ref: https://github.com/scxue/SA-Solver import math from typing import Union, Callable import torch def compute_exponential_coeffs(s: torch.Tensor, t: torch.Tensor, solver_order: int, tau_t: float) -> torch.Tensor: """Compute (1 + tau^2) * integral of exp((1 + tau^2) * x) * x^p dx from s to t with exp((1 + tau^2) * t) factored out, using integration by parts. Integral of exp((1 + tau^2) * x) * x^p dx = product_terms[p] - (p / (1 + tau^2)) * integral of exp((1 + tau^2) * x) * x^(p-1) dx, with base case p=0 where integral equals product_terms[0]. where product_terms[p] = x^p * exp((1 + tau^2) * x) / (1 + tau^2). Construct a recursive coefficient matrix following the above recursive relation to compute all integral terms up to p = (solver_order - 1). Return coefficients used by the SA-Solver in data prediction mode. Args: s: Start time s. t: End time t. solver_order: Current order of the solver. tau_t: Stochastic strength parameter in the SDE. Returns: Exponential coefficients used in data prediction, with exp((1 + tau^2) * t) factored out, ordered from p=0 to p=solver_order−1, shape (solver_order,). """ tau_mul = 1 + tau_t ** 2 h = t - s p = torch.arange(solver_order, dtype=s.dtype, device=s.device) # product_terms after factoring out exp((1 + tau^2) * t) # Includes (1 + tau^2) factor from outside the integral product_terms_factored = (t ** p - s ** p * (-tau_mul * h).exp()) # Lower triangular recursive coefficient matrix # Accumulates recursive coefficients based on p / (1 + tau^2) recursive_depth_mat = p.unsqueeze(1) - p.unsqueeze(0) log_factorial = (p + 1).lgamma() recursive_coeff_mat = log_factorial.unsqueeze(1) - log_factorial.unsqueeze(0) if tau_t > 0: recursive_coeff_mat = recursive_coeff_mat - (recursive_depth_mat * math.log(tau_mul)) signs = torch.where(recursive_depth_mat % 2 == 0, 1.0, -1.0) recursive_coeff_mat = (recursive_coeff_mat.exp() * signs).tril() return recursive_coeff_mat @ product_terms_factored def compute_simple_stochastic_adams_b_coeffs(sigma_next: torch.Tensor, curr_lambdas: torch.Tensor, lambda_s: torch.Tensor, lambda_t: torch.Tensor, tau_t: float, is_corrector_step: bool = False) -> torch.Tensor: """Compute simple order-2 b coefficients from SA-Solver paper (Appendix D. Implementation Details).""" tau_mul = 1 + tau_t ** 2 h = lambda_t - lambda_s alpha_t = sigma_next * lambda_t.exp() if is_corrector_step: # Simplified 1-step (order-2) corrector b_1 = alpha_t * (0.5 * tau_mul * h) b_2 = alpha_t * (-h * tau_mul).expm1().neg() - b_1 else: # Simplified 2-step predictor b_2 = alpha_t * (0.5 * tau_mul * h ** 2) / (curr_lambdas[-2] - lambda_s) b_1 = alpha_t * (-h * tau_mul).expm1().neg() - b_2 return torch.stack([b_2, b_1]) def compute_stochastic_adams_b_coeffs(sigma_next: torch.Tensor, curr_lambdas: torch.Tensor, lambda_s: torch.Tensor, lambda_t: torch.Tensor, tau_t: float, simple_order_2: bool = False, is_corrector_step: bool = False) -> torch.Tensor: """Compute b_i coefficients for the SA-Solver (see eqs. 15 and 18). The solver order corresponds to the number of input lambdas (half-logSNR points). Args: sigma_next: Sigma at end time t. curr_lambdas: Lambda time points used to construct the Lagrange basis, shape (N,). lambda_s: Lambda at start time s. lambda_t: Lambda at end time t. tau_t: Stochastic strength parameter in the SDE. simple_order_2: Whether to enable the simple order-2 scheme. is_corrector_step: Flag for corrector step in simple order-2 mode. Returns: b_i coefficients for the SA-Solver, shape (N,), where N is the solver order. """ num_timesteps = curr_lambdas.shape[0] if simple_order_2 and num_timesteps == 2: return compute_simple_stochastic_adams_b_coeffs(sigma_next, curr_lambdas, lambda_s, lambda_t, tau_t, is_corrector_step) # Compute coefficients by solving a linear system from Lagrange basis interpolation exp_integral_coeffs = compute_exponential_coeffs(lambda_s, lambda_t, num_timesteps, tau_t) vandermonde_matrix_T = torch.vander(curr_lambdas, num_timesteps, increasing=True).T lagrange_integrals = torch.linalg.solve(vandermonde_matrix_T, exp_integral_coeffs) # (sigma_t * exp(-tau^2 * lambda_t)) * exp((1 + tau^2) * lambda_t) # = sigma_t * exp(lambda_t) = alpha_t # exp((1 + tau^2) * lambda_t) is extracted from the integral alpha_t = sigma_next * lambda_t.exp() return alpha_t * lagrange_integrals def get_tau_interval_func(start_sigma: float, end_sigma: float, eta: float = 1.0) -> Callable[[Union[torch.Tensor, float]], float]: """Return a function that controls the stochasticity of SA-Solver. When eta = 0, SA-Solver runs as ODE. The official approach uses time t to determine the SDE interval, while here we use sigma instead. See: https://github.com/scxue/SA-Solver/blob/main/README.md """ def tau_func(sigma: Union[torch.Tensor, float]) -> float: if eta <= 0: return 0.0 # ODE if isinstance(sigma, torch.Tensor): sigma = sigma.item() return eta if start_sigma >= sigma >= end_sigma else 0.0 return tau_func
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/k_diffusion/sa_solver.py", "license": "GNU General Public License v3.0", "lines": 93, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_api_nodes/nodes_moonvalley.py
import logging from typing_extensions import override from comfy_api.latest import IO, ComfyExtension, Input from comfy_api_nodes.apis.moonvalley import ( MoonvalleyPromptResponse, MoonvalleyTextToVideoInferenceParams, MoonvalleyTextToVideoRequest, MoonvalleyVideoToVideoInferenceParams, MoonvalleyVideoToVideoRequest, ) from comfy_api_nodes.util import ( ApiEndpoint, download_url_to_video_output, poll_op, sync_op, trim_video, upload_images_to_comfyapi, upload_video_to_comfyapi, validate_container_format_is_mp4, validate_image_dimensions, validate_string, ) API_UPLOADS_ENDPOINT = "/proxy/moonvalley/uploads" API_PROMPTS_ENDPOINT = "/proxy/moonvalley/prompts" API_VIDEO2VIDEO_ENDPOINT = "/proxy/moonvalley/prompts/video-to-video" API_TXT2VIDEO_ENDPOINT = "/proxy/moonvalley/prompts/text-to-video" API_IMG2VIDEO_ENDPOINT = "/proxy/moonvalley/prompts/image-to-video" MIN_WIDTH = 300 MIN_HEIGHT = 300 MAX_WIDTH = 10000 MAX_HEIGHT = 10000 MIN_VID_WIDTH = 300 MIN_VID_HEIGHT = 300 MAX_VID_WIDTH = 10000 MAX_VID_HEIGHT = 10000 MAX_VIDEO_SIZE = 1024 * 1024 * 1024 # 1 GB max for in-memory video processing MOONVALLEY_MAREY_MAX_PROMPT_LENGTH = 5000 def is_valid_task_creation_response(response: MoonvalleyPromptResponse) -> bool: """Verifies that the initial response contains a task ID.""" return bool(response.id) def validate_task_creation_response(response) -> None: if not is_valid_task_creation_response(response): error_msg = f"Moonvalley Marey API: Initial request failed. Code: {response.code}, Message: {response.message}, Data: {response}" logging.error(error_msg) raise RuntimeError(error_msg) def validate_video_to_video_input(video: Input.Video) -> Input.Video: """ Validates and processes video input for Moonvalley Video-to-Video generation. Args: video: Input video to validate Returns: Validated and potentially trimmed video Raises: ValueError: If video doesn't meet requirements MoonvalleyApiError: If video duration is too short """ width, height = _get_video_dimensions(video) _validate_video_dimensions(width, height) validate_container_format_is_mp4(video) return _validate_and_trim_duration(video) def _get_video_dimensions(video: Input.Video) -> tuple[int, int]: """Extracts video dimensions with error handling.""" try: return video.get_dimensions() except Exception as e: logging.error("Error getting dimensions of video: %s", e) raise ValueError(f"Cannot get video dimensions: {e}") from e def _validate_video_dimensions(width: int, height: int) -> None: """Validates video dimensions meet Moonvalley V2V requirements.""" supported_resolutions = { (1920, 1080), (1080, 1920), (1152, 1152), (1536, 1152), (1152, 1536), } if (width, height) not in supported_resolutions: supported_list = ", ".join([f"{w}x{h}" for w, h in sorted(supported_resolutions)]) raise ValueError(f"Resolution {width}x{height} not supported. Supported: {supported_list}") def _validate_and_trim_duration(video: Input.Video) -> Input.Video: """Validates video duration and trims to 5 seconds if needed.""" duration = video.get_duration() _validate_minimum_duration(duration) return _trim_if_too_long(video, duration) def _validate_minimum_duration(duration: float) -> None: """Ensures video is at least 5 seconds long.""" if duration < 5: raise ValueError("Input video must be at least 5 seconds long.") def _trim_if_too_long(video: Input.Video, duration: float) -> Input.Video: """Trims video to 5 seconds if longer.""" if duration > 5: return trim_video(video, 5) return video def parse_width_height_from_res(resolution: str): # Accepts a string like "16:9 (1920 x 1080)" and returns width, height as a dict res_map = { "16:9 (1920 x 1080)": {"width": 1920, "height": 1080}, "9:16 (1080 x 1920)": {"width": 1080, "height": 1920}, "1:1 (1152 x 1152)": {"width": 1152, "height": 1152}, "4:3 (1536 x 1152)": {"width": 1536, "height": 1152}, "3:4 (1152 x 1536)": {"width": 1152, "height": 1536}, # "21:9 (2560 x 1080)": {"width": 2560, "height": 1080}, } return res_map.get(resolution, {"width": 1920, "height": 1080}) def parse_control_parameter(value): control_map = { "Motion Transfer": "motion_control", "Canny": "canny_control", "Pose Transfer": "pose_control", "Depth": "depth_control", } return control_map.get(value, control_map["Motion Transfer"]) async def get_response(cls: type[IO.ComfyNode], task_id: str) -> MoonvalleyPromptResponse: return await poll_op( cls, ApiEndpoint(path=f"{API_PROMPTS_ENDPOINT}/{task_id}"), response_model=MoonvalleyPromptResponse, status_extractor=lambda r: (r.status if r and r.status else None), poll_interval=16.0, max_poll_attempts=240, ) class MoonvalleyImg2VideoNode(IO.ComfyNode): @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id="MoonvalleyImg2VideoNode", display_name="Moonvalley Marey Image to Video", category="api node/video/Moonvalley Marey", description="Moonvalley Marey Image to Video Node", inputs=[ IO.Image.Input( "image", tooltip="The reference image used to generate the video", ), IO.String.Input( "prompt", multiline=True, ), IO.String.Input( "negative_prompt", multiline=True, default="<synthetic> <scene cut> gopro, bright, contrast, static, overexposed, vignette, " "artifacts, still, noise, texture, scanlines, videogame, 360 camera, VR, transition, " "flare, saturation, distorted, warped, wide angle, saturated, vibrant, glowing, " "cross dissolve, cheesy, ugly hands, mutated hands, mutant, disfigured, extra fingers, " "blown out, horrible, blurry, worst quality, bad, dissolve, melt, fade in, fade out, " "wobbly, weird, low quality, plastic, stock footage, video camera, boring", tooltip="Negative prompt text", ), IO.Combo.Input( "resolution", options=[ "16:9 (1920 x 1080)", "9:16 (1080 x 1920)", "1:1 (1152 x 1152)", "4:3 (1536 x 1152)", "3:4 (1152 x 1536)", # "21:9 (2560 x 1080)", ], default="16:9 (1920 x 1080)", tooltip="Resolution of the output video", ), IO.Float.Input( "prompt_adherence", default=4.5, min=1.0, max=20.0, step=1.0, tooltip="Guidance scale for generation control", ), IO.Int.Input( "seed", default=9, min=0, max=4294967295, step=1, display_mode=IO.NumberDisplay.number, tooltip="Random seed value", control_after_generate=True, ), IO.Int.Input( "steps", default=80, min=75, # steps should be greater or equal to cooldown_steps(75) + warmup_steps(0) max=100, step=1, tooltip="Number of denoising steps", ), ], outputs=[IO.Video.Output()], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(), expr="""{"type":"usd","usd": 1.5}""", ), ) @classmethod async def execute( cls, image: Input.Image, prompt: str, negative_prompt: str, resolution: str, prompt_adherence: float, seed: int, steps: int, ) -> IO.NodeOutput: validate_image_dimensions(image, min_width=300, min_height=300, max_height=MAX_HEIGHT, max_width=MAX_WIDTH) validate_string(prompt, min_length=1, max_length=MOONVALLEY_MAREY_MAX_PROMPT_LENGTH) validate_string(negative_prompt, field_name="negative_prompt", max_length=MOONVALLEY_MAREY_MAX_PROMPT_LENGTH) width_height = parse_width_height_from_res(resolution) inference_params = MoonvalleyTextToVideoInferenceParams( negative_prompt=negative_prompt, steps=steps, seed=seed, guidance_scale=prompt_adherence, width=width_height["width"], height=width_height["height"], use_negative_prompts=True, ) # Get MIME type from tensor - assuming PNG format for image tensors mime_type = "image/png" image_url = (await upload_images_to_comfyapi(cls, image, max_images=1, mime_type=mime_type))[0] task_creation_response = await sync_op( cls, endpoint=ApiEndpoint(path=API_IMG2VIDEO_ENDPOINT, method="POST"), response_model=MoonvalleyPromptResponse, data=MoonvalleyTextToVideoRequest( image_url=image_url, prompt_text=prompt, inference_params=inference_params ), ) validate_task_creation_response(task_creation_response) final_response = await get_response(cls, task_creation_response.id) video = await download_url_to_video_output(final_response.output_url) return IO.NodeOutput(video) class MoonvalleyVideo2VideoNode(IO.ComfyNode): @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id="MoonvalleyVideo2VideoNode", display_name="Moonvalley Marey Video to Video", category="api node/video/Moonvalley Marey", description="", inputs=[ IO.String.Input( "prompt", multiline=True, tooltip="Describes the video to generate", ), IO.String.Input( "negative_prompt", multiline=True, default="<synthetic> <scene cut> gopro, bright, contrast, static, overexposed, vignette, " "artifacts, still, noise, texture, scanlines, videogame, 360 camera, VR, transition, " "flare, saturation, distorted, warped, wide angle, saturated, vibrant, glowing, " "cross dissolve, cheesy, ugly hands, mutated hands, mutant, disfigured, extra fingers, " "blown out, horrible, blurry, worst quality, bad, dissolve, melt, fade in, fade out, " "wobbly, weird, low quality, plastic, stock footage, video camera, boring", tooltip="Negative prompt text", ), IO.Int.Input( "seed", default=9, min=0, max=4294967295, step=1, display_mode=IO.NumberDisplay.number, tooltip="Random seed value", control_after_generate=False, ), IO.Video.Input( "video", tooltip="The reference video used to generate the output video. Must be at least 5 seconds long. " "Videos longer than 5s will be automatically trimmed. Only MP4 format supported.", ), IO.Combo.Input( "control_type", options=["Motion Transfer", "Pose Transfer"], default="Motion Transfer", optional=True, ), IO.Int.Input( "motion_intensity", default=100, min=0, max=100, step=1, tooltip="Only used if control_type is 'Motion Transfer'", optional=True, ), IO.Int.Input( "steps", default=60, min=60, # steps should be greater or equal to cooldown_steps(36) + warmup_steps(24) max=100, step=1, display_mode=IO.NumberDisplay.number, tooltip="Number of inference steps", ), ], outputs=[IO.Video.Output()], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(), expr="""{"type":"usd","usd": 2.25}""", ), ) @classmethod async def execute( cls, prompt: str, negative_prompt: str, seed: int, video: Input.Video | None = None, control_type: str = "Motion Transfer", motion_intensity: int | None = 100, steps=60, prompt_adherence=4.5, ) -> IO.NodeOutput: validated_video = validate_video_to_video_input(video) video_url = await upload_video_to_comfyapi(cls, validated_video) validate_string(prompt, min_length=1, max_length=MOONVALLEY_MAREY_MAX_PROMPT_LENGTH) validate_string(negative_prompt, field_name="negative_prompt", max_length=MOONVALLEY_MAREY_MAX_PROMPT_LENGTH) # Only include motion_intensity for Motion Transfer control_params = {} if control_type == "Motion Transfer" and motion_intensity is not None: control_params["motion_intensity"] = motion_intensity inference_params = MoonvalleyVideoToVideoInferenceParams( negative_prompt=negative_prompt, seed=seed, control_params=control_params, steps=steps, guidance_scale=prompt_adherence, ) task_creation_response = await sync_op( cls, endpoint=ApiEndpoint(path=API_VIDEO2VIDEO_ENDPOINT, method="POST"), response_model=MoonvalleyPromptResponse, data=MoonvalleyVideoToVideoRequest( control_type=parse_control_parameter(control_type), video_url=video_url, prompt_text=prompt, inference_params=inference_params, ), ) validate_task_creation_response(task_creation_response) final_response = await get_response(cls, task_creation_response.id) return IO.NodeOutput(await download_url_to_video_output(final_response.output_url)) class MoonvalleyTxt2VideoNode(IO.ComfyNode): @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id="MoonvalleyTxt2VideoNode", display_name="Moonvalley Marey Text to Video", category="api node/video/Moonvalley Marey", description="", inputs=[ IO.String.Input( "prompt", multiline=True, ), IO.String.Input( "negative_prompt", multiline=True, default="<synthetic> <scene cut> gopro, bright, contrast, static, overexposed, vignette, " "artifacts, still, noise, texture, scanlines, videogame, 360 camera, VR, transition, " "flare, saturation, distorted, warped, wide angle, saturated, vibrant, glowing, " "cross dissolve, cheesy, ugly hands, mutated hands, mutant, disfigured, extra fingers, " "blown out, horrible, blurry, worst quality, bad, dissolve, melt, fade in, fade out, " "wobbly, weird, low quality, plastic, stock footage, video camera, boring", tooltip="Negative prompt text", ), IO.Combo.Input( "resolution", options=[ "16:9 (1920 x 1080)", "9:16 (1080 x 1920)", "1:1 (1152 x 1152)", "4:3 (1536 x 1152)", "3:4 (1152 x 1536)", "21:9 (2560 x 1080)", ], default="16:9 (1920 x 1080)", tooltip="Resolution of the output video", ), IO.Float.Input( "prompt_adherence", default=4.0, min=1.0, max=20.0, step=1.0, tooltip="Guidance scale for generation control", ), IO.Int.Input( "seed", default=9, min=0, max=4294967295, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, tooltip="Random seed value", ), IO.Int.Input( "steps", default=80, min=75, # steps should be greater or equal to cooldown_steps(75) + warmup_steps(0) max=100, step=1, tooltip="Inference steps", ), ], outputs=[IO.Video.Output()], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(), expr="""{"type":"usd","usd": 1.5}""", ), ) @classmethod async def execute( cls, prompt: str, negative_prompt: str, resolution: str, prompt_adherence: float, seed: int, steps: int, ) -> IO.NodeOutput: validate_string(prompt, min_length=1, max_length=MOONVALLEY_MAREY_MAX_PROMPT_LENGTH) validate_string(negative_prompt, field_name="negative_prompt", max_length=MOONVALLEY_MAREY_MAX_PROMPT_LENGTH) width_height = parse_width_height_from_res(resolution) inference_params = MoonvalleyTextToVideoInferenceParams( negative_prompt=negative_prompt, steps=steps, seed=seed, guidance_scale=prompt_adherence, num_frames=128, width=width_height["width"], height=width_height["height"], ) task_creation_response = await sync_op( cls, endpoint=ApiEndpoint(path=API_TXT2VIDEO_ENDPOINT, method="POST"), response_model=MoonvalleyPromptResponse, data=MoonvalleyTextToVideoRequest(prompt_text=prompt, inference_params=inference_params), ) validate_task_creation_response(task_creation_response) final_response = await get_response(cls, task_creation_response.id) return IO.NodeOutput(await download_url_to_video_output(final_response.output_url)) class MoonvalleyExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[IO.ComfyNode]]: return [ MoonvalleyImg2VideoNode, MoonvalleyTxt2VideoNode, MoonvalleyVideo2VideoNode, ] async def comfy_entrypoint() -> MoonvalleyExtension: return MoonvalleyExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api_nodes/nodes_moonvalley.py", "license": "GNU General Public License v3.0", "lines": 475, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_extras/nodes_tcfg.py
# TCFG: Tangential Damping Classifier-free Guidance - (arXiv: https://arxiv.org/abs/2503.18137) from typing_extensions import override import torch from comfy_api.latest import ComfyExtension, io def score_tangential_damping(cond_score: torch.Tensor, uncond_score: torch.Tensor) -> torch.Tensor: """Drop tangential components from uncond score to align with cond score.""" # (B, 1, ...) batch_num = cond_score.shape[0] cond_score_flat = cond_score.reshape(batch_num, 1, -1).float() uncond_score_flat = uncond_score.reshape(batch_num, 1, -1).float() # Score matrix A (B, 2, ...) score_matrix = torch.cat((uncond_score_flat, cond_score_flat), dim=1) try: _, _, Vh = torch.linalg.svd(score_matrix, full_matrices=False) except RuntimeError: # Fallback to CPU _, _, Vh = torch.linalg.svd(score_matrix.cpu(), full_matrices=False) # Drop the tangential components v1 = Vh[:, 0:1, :].to(uncond_score_flat.device) # (B, 1, ...) uncond_score_td = (uncond_score_flat @ v1.transpose(-2, -1)) * v1 return uncond_score_td.reshape_as(uncond_score).to(uncond_score.dtype) class TCFG(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="TCFG", display_name="Tangential Damping CFG", category="advanced/guidance", description="TCFG – Tangential Damping CFG (2503.18137)\n\nRefine the uncond (negative) to align with the cond (positive) for improving quality.", inputs=[ io.Model.Input("model"), ], outputs=[ io.Model.Output(display_name="patched_model"), ], ) @classmethod def execute(cls, model): m = model.clone() def tangential_damping_cfg(args): # Assume [cond, uncond, ...] x = args["input"] conds_out = args["conds_out"] if len(conds_out) <= 1 or None in args["conds"][:2]: # Skip when either cond or uncond is None return conds_out cond_pred = conds_out[0] uncond_pred = conds_out[1] uncond_td = score_tangential_damping(x - cond_pred, x - uncond_pred) uncond_pred_td = x - uncond_td return [cond_pred, uncond_pred_td] + conds_out[2:] m.set_model_sampler_pre_cfg_function(tangential_damping_cfg) return io.NodeOutput(m) class TcfgExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[io.ComfyNode]]: return [ TCFG, ] async def comfy_entrypoint() -> TcfgExtension: return TcfgExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_extras/nodes_tcfg.py", "license": "GNU General Public License v3.0", "lines": 61, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy/ldm/omnigen/omnigen2.py
# Original code: https://github.com/VectorSpaceLab/OmniGen2 from typing import Optional, Tuple import torch import torch.nn as nn import torch.nn.functional as F from einops import rearrange, repeat from comfy.ldm.lightricks.model import Timesteps from comfy.ldm.flux.layers import EmbedND from comfy.ldm.modules.attention import optimized_attention_masked import comfy.model_management import comfy.ldm.common_dit def apply_rotary_emb(x, freqs_cis): if x.shape[1] == 0: return x t_ = x.reshape(*x.shape[:-1], -1, 1, 2) t_out = freqs_cis[..., 0] * t_[..., 0] + freqs_cis[..., 1] * t_[..., 1] return t_out.reshape(*x.shape).to(dtype=x.dtype) def swiglu(x: torch.Tensor, y: torch.Tensor) -> torch.Tensor: return F.silu(x) * y class TimestepEmbedding(nn.Module): def __init__(self, in_channels: int, time_embed_dim: int, dtype=None, device=None, operations=None): super().__init__() self.linear_1 = operations.Linear(in_channels, time_embed_dim, dtype=dtype, device=device) self.act = nn.SiLU() self.linear_2 = operations.Linear(time_embed_dim, time_embed_dim, dtype=dtype, device=device) def forward(self, sample: torch.Tensor) -> torch.Tensor: sample = self.linear_1(sample) sample = self.act(sample) sample = self.linear_2(sample) return sample class LuminaRMSNormZero(nn.Module): def __init__(self, embedding_dim: int, norm_eps: float = 1e-5, dtype=None, device=None, operations=None): super().__init__() self.silu = nn.SiLU() self.linear = operations.Linear(min(embedding_dim, 1024), 4 * embedding_dim, dtype=dtype, device=device) self.norm = operations.RMSNorm(embedding_dim, eps=norm_eps, dtype=dtype, device=device) def forward(self, x: torch.Tensor, emb: torch.Tensor) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor]: emb = self.linear(self.silu(emb)) scale_msa, gate_msa, scale_mlp, gate_mlp = emb.chunk(4, dim=1) x = self.norm(x) * (1 + scale_msa[:, None]) return x, gate_msa, scale_mlp, gate_mlp class LuminaLayerNormContinuous(nn.Module): def __init__(self, embedding_dim: int, conditioning_embedding_dim: int, elementwise_affine: bool = False, eps: float = 1e-6, out_dim: Optional[int] = None, dtype=None, device=None, operations=None): super().__init__() self.silu = nn.SiLU() self.linear_1 = operations.Linear(conditioning_embedding_dim, embedding_dim, dtype=dtype, device=device) self.norm = operations.LayerNorm(embedding_dim, eps, elementwise_affine, dtype=dtype, device=device) self.linear_2 = operations.Linear(embedding_dim, out_dim, bias=True, dtype=dtype, device=device) if out_dim is not None else None def forward(self, x: torch.Tensor, conditioning_embedding: torch.Tensor) -> torch.Tensor: emb = self.linear_1(self.silu(conditioning_embedding).to(x.dtype)) x = self.norm(x) * (1 + emb)[:, None, :] if self.linear_2 is not None: x = self.linear_2(x) return x class LuminaFeedForward(nn.Module): def __init__(self, dim: int, inner_dim: int, multiple_of: int = 256, dtype=None, device=None, operations=None): super().__init__() inner_dim = multiple_of * ((inner_dim + multiple_of - 1) // multiple_of) self.linear_1 = operations.Linear(dim, inner_dim, bias=False, dtype=dtype, device=device) self.linear_2 = operations.Linear(inner_dim, dim, bias=False, dtype=dtype, device=device) self.linear_3 = operations.Linear(dim, inner_dim, bias=False, dtype=dtype, device=device) def forward(self, x: torch.Tensor) -> torch.Tensor: h1, h2 = self.linear_1(x), self.linear_3(x) return self.linear_2(swiglu(h1, h2)) class Lumina2CombinedTimestepCaptionEmbedding(nn.Module): def __init__(self, hidden_size: int = 4096, text_feat_dim: int = 2048, frequency_embedding_size: int = 256, norm_eps: float = 1e-5, timestep_scale: float = 1.0, dtype=None, device=None, operations=None): super().__init__() self.time_proj = Timesteps(num_channels=frequency_embedding_size, flip_sin_to_cos=True, downscale_freq_shift=0.0, scale=timestep_scale) self.timestep_embedder = TimestepEmbedding(in_channels=frequency_embedding_size, time_embed_dim=min(hidden_size, 1024), dtype=dtype, device=device, operations=operations) self.caption_embedder = nn.Sequential( operations.RMSNorm(text_feat_dim, eps=norm_eps, dtype=dtype, device=device), operations.Linear(text_feat_dim, hidden_size, bias=True, dtype=dtype, device=device), ) def forward(self, timestep: torch.Tensor, text_hidden_states: torch.Tensor, dtype: torch.dtype) -> Tuple[torch.Tensor, torch.Tensor]: timestep_proj = self.time_proj(timestep).to(dtype=dtype) time_embed = self.timestep_embedder(timestep_proj) caption_embed = self.caption_embedder(text_hidden_states) return time_embed, caption_embed class Attention(nn.Module): def __init__(self, query_dim: int, dim_head: int, heads: int, kv_heads: int, eps: float = 1e-5, bias: bool = False, dtype=None, device=None, operations=None): super().__init__() self.heads = heads self.kv_heads = kv_heads self.dim_head = dim_head self.scale = dim_head ** -0.5 self.to_q = operations.Linear(query_dim, heads * dim_head, bias=bias, dtype=dtype, device=device) self.to_k = operations.Linear(query_dim, kv_heads * dim_head, bias=bias, dtype=dtype, device=device) self.to_v = operations.Linear(query_dim, kv_heads * dim_head, bias=bias, dtype=dtype, device=device) self.norm_q = operations.RMSNorm(dim_head, eps=eps, dtype=dtype, device=device) self.norm_k = operations.RMSNorm(dim_head, eps=eps, dtype=dtype, device=device) self.to_out = nn.Sequential( operations.Linear(heads * dim_head, query_dim, bias=bias, dtype=dtype, device=device), nn.Dropout(0.0) ) def forward(self, hidden_states: torch.Tensor, encoder_hidden_states: torch.Tensor, attention_mask: Optional[torch.Tensor] = None, image_rotary_emb: Optional[torch.Tensor] = None, transformer_options={}) -> torch.Tensor: batch_size, sequence_length, _ = hidden_states.shape query = self.to_q(hidden_states) key = self.to_k(encoder_hidden_states) value = self.to_v(encoder_hidden_states) query = query.view(batch_size, -1, self.heads, self.dim_head) key = key.view(batch_size, -1, self.kv_heads, self.dim_head) value = value.view(batch_size, -1, self.kv_heads, self.dim_head) query = self.norm_q(query) key = self.norm_k(key) if image_rotary_emb is not None: query = apply_rotary_emb(query, image_rotary_emb) key = apply_rotary_emb(key, image_rotary_emb) query = query.transpose(1, 2) key = key.transpose(1, 2) value = value.transpose(1, 2) if self.kv_heads < self.heads: key = key.repeat_interleave(self.heads // self.kv_heads, dim=1) value = value.repeat_interleave(self.heads // self.kv_heads, dim=1) hidden_states = optimized_attention_masked(query, key, value, self.heads, attention_mask, skip_reshape=True, transformer_options=transformer_options) hidden_states = self.to_out[0](hidden_states) return hidden_states class OmniGen2TransformerBlock(nn.Module): def __init__(self, dim: int, num_attention_heads: int, num_kv_heads: int, multiple_of: int, ffn_dim_multiplier: float, norm_eps: float, modulation: bool = True, dtype=None, device=None, operations=None): super().__init__() self.modulation = modulation self.attn = Attention( query_dim=dim, dim_head=dim // num_attention_heads, heads=num_attention_heads, kv_heads=num_kv_heads, eps=1e-5, bias=False, dtype=dtype, device=device, operations=operations, ) self.feed_forward = LuminaFeedForward( dim=dim, inner_dim=4 * dim, multiple_of=multiple_of, dtype=dtype, device=device, operations=operations ) if modulation: self.norm1 = LuminaRMSNormZero(embedding_dim=dim, norm_eps=norm_eps, dtype=dtype, device=device, operations=operations) else: self.norm1 = operations.RMSNorm(dim, eps=norm_eps, dtype=dtype, device=device) self.ffn_norm1 = operations.RMSNorm(dim, eps=norm_eps, dtype=dtype, device=device) self.norm2 = operations.RMSNorm(dim, eps=norm_eps, dtype=dtype, device=device) self.ffn_norm2 = operations.RMSNorm(dim, eps=norm_eps, dtype=dtype, device=device) def forward(self, hidden_states: torch.Tensor, attention_mask: torch.Tensor, image_rotary_emb: torch.Tensor, temb: Optional[torch.Tensor] = None, transformer_options={}) -> torch.Tensor: if self.modulation: norm_hidden_states, gate_msa, scale_mlp, gate_mlp = self.norm1(hidden_states, temb) attn_output = self.attn(norm_hidden_states, norm_hidden_states, attention_mask, image_rotary_emb, transformer_options=transformer_options) hidden_states = hidden_states + gate_msa.unsqueeze(1).tanh() * self.norm2(attn_output) mlp_output = self.feed_forward(self.ffn_norm1(hidden_states) * (1 + scale_mlp.unsqueeze(1))) hidden_states = hidden_states + gate_mlp.unsqueeze(1).tanh() * self.ffn_norm2(mlp_output) else: norm_hidden_states = self.norm1(hidden_states) attn_output = self.attn(norm_hidden_states, norm_hidden_states, attention_mask, image_rotary_emb, transformer_options=transformer_options) hidden_states = hidden_states + self.norm2(attn_output) mlp_output = self.feed_forward(self.ffn_norm1(hidden_states)) hidden_states = hidden_states + self.ffn_norm2(mlp_output) return hidden_states class OmniGen2RotaryPosEmbed(nn.Module): def __init__(self, theta: int, axes_dim: Tuple[int, int, int], axes_lens: Tuple[int, int, int] = (300, 512, 512), patch_size: int = 2): super().__init__() self.theta = theta self.axes_dim = axes_dim self.axes_lens = axes_lens self.patch_size = patch_size self.rope_embedder = EmbedND(dim=sum(axes_dim), theta=self.theta, axes_dim=axes_dim) def forward(self, batch_size, encoder_seq_len, l_effective_cap_len, l_effective_ref_img_len, l_effective_img_len, ref_img_sizes, img_sizes, device): p = self.patch_size seq_lengths = [cap_len + sum(ref_img_len) + img_len for cap_len, ref_img_len, img_len in zip(l_effective_cap_len, l_effective_ref_img_len, l_effective_img_len)] max_seq_len = max(seq_lengths) max_ref_img_len = max([sum(ref_img_len) for ref_img_len in l_effective_ref_img_len]) max_img_len = max(l_effective_img_len) position_ids = torch.zeros(batch_size, max_seq_len, 3, dtype=torch.int32, device=device) for i, (cap_seq_len, seq_len) in enumerate(zip(l_effective_cap_len, seq_lengths)): position_ids[i, :cap_seq_len] = repeat(torch.arange(cap_seq_len, dtype=torch.int32, device=device), "l -> l 3") pe_shift = cap_seq_len pe_shift_len = cap_seq_len if ref_img_sizes[i] is not None: for ref_img_size, ref_img_len in zip(ref_img_sizes[i], l_effective_ref_img_len[i]): H, W = ref_img_size ref_H_tokens, ref_W_tokens = H // p, W // p row_ids = repeat(torch.arange(ref_H_tokens, dtype=torch.int32, device=device), "h -> h w", w=ref_W_tokens).flatten() col_ids = repeat(torch.arange(ref_W_tokens, dtype=torch.int32, device=device), "w -> h w", h=ref_H_tokens).flatten() position_ids[i, pe_shift_len:pe_shift_len + ref_img_len, 0] = pe_shift position_ids[i, pe_shift_len:pe_shift_len + ref_img_len, 1] = row_ids position_ids[i, pe_shift_len:pe_shift_len + ref_img_len, 2] = col_ids pe_shift += max(ref_H_tokens, ref_W_tokens) pe_shift_len += ref_img_len H, W = img_sizes[i] H_tokens, W_tokens = H // p, W // p row_ids = repeat(torch.arange(H_tokens, dtype=torch.int32, device=device), "h -> h w", w=W_tokens).flatten() col_ids = repeat(torch.arange(W_tokens, dtype=torch.int32, device=device), "w -> h w", h=H_tokens).flatten() position_ids[i, pe_shift_len: seq_len, 0] = pe_shift position_ids[i, pe_shift_len: seq_len, 1] = row_ids position_ids[i, pe_shift_len: seq_len, 2] = col_ids freqs_cis = self.rope_embedder(position_ids).movedim(1, 2) cap_freqs_cis_shape = list(freqs_cis.shape) cap_freqs_cis_shape[1] = encoder_seq_len cap_freqs_cis = torch.zeros(*cap_freqs_cis_shape, device=device, dtype=freqs_cis.dtype) ref_img_freqs_cis_shape = list(freqs_cis.shape) ref_img_freqs_cis_shape[1] = max_ref_img_len ref_img_freqs_cis = torch.zeros(*ref_img_freqs_cis_shape, device=device, dtype=freqs_cis.dtype) img_freqs_cis_shape = list(freqs_cis.shape) img_freqs_cis_shape[1] = max_img_len img_freqs_cis = torch.zeros(*img_freqs_cis_shape, device=device, dtype=freqs_cis.dtype) for i, (cap_seq_len, ref_img_len, img_len, seq_len) in enumerate(zip(l_effective_cap_len, l_effective_ref_img_len, l_effective_img_len, seq_lengths)): cap_freqs_cis[i, :cap_seq_len] = freqs_cis[i, :cap_seq_len] ref_img_freqs_cis[i, :sum(ref_img_len)] = freqs_cis[i, cap_seq_len:cap_seq_len + sum(ref_img_len)] img_freqs_cis[i, :img_len] = freqs_cis[i, cap_seq_len + sum(ref_img_len):cap_seq_len + sum(ref_img_len) + img_len] return cap_freqs_cis, ref_img_freqs_cis, img_freqs_cis, freqs_cis, l_effective_cap_len, seq_lengths class OmniGen2Transformer2DModel(nn.Module): def __init__( self, patch_size: int = 2, in_channels: int = 16, out_channels: Optional[int] = None, hidden_size: int = 2304, num_layers: int = 26, num_refiner_layers: int = 2, num_attention_heads: int = 24, num_kv_heads: int = 8, multiple_of: int = 256, ffn_dim_multiplier: Optional[float] = None, norm_eps: float = 1e-5, axes_dim_rope: Tuple[int, int, int] = (32, 32, 32), axes_lens: Tuple[int, int, int] = (300, 512, 512), text_feat_dim: int = 1024, timestep_scale: float = 1.0, image_model=None, device=None, dtype=None, operations=None, ): super().__init__() self.patch_size = patch_size self.out_channels = out_channels or in_channels self.hidden_size = hidden_size self.dtype = dtype self.rope_embedder = OmniGen2RotaryPosEmbed( theta=10000, axes_dim=axes_dim_rope, axes_lens=axes_lens, patch_size=patch_size, ) self.x_embedder = operations.Linear(patch_size * patch_size * in_channels, hidden_size, dtype=dtype, device=device) self.ref_image_patch_embedder = operations.Linear(patch_size * patch_size * in_channels, hidden_size, dtype=dtype, device=device) self.time_caption_embed = Lumina2CombinedTimestepCaptionEmbedding( hidden_size=hidden_size, text_feat_dim=text_feat_dim, norm_eps=norm_eps, timestep_scale=timestep_scale, dtype=dtype, device=device, operations=operations ) self.noise_refiner = nn.ModuleList([ OmniGen2TransformerBlock( hidden_size, num_attention_heads, num_kv_heads, multiple_of, ffn_dim_multiplier, norm_eps, modulation=True, dtype=dtype, device=device, operations=operations ) for _ in range(num_refiner_layers) ]) self.ref_image_refiner = nn.ModuleList([ OmniGen2TransformerBlock( hidden_size, num_attention_heads, num_kv_heads, multiple_of, ffn_dim_multiplier, norm_eps, modulation=True, dtype=dtype, device=device, operations=operations ) for _ in range(num_refiner_layers) ]) self.context_refiner = nn.ModuleList([ OmniGen2TransformerBlock( hidden_size, num_attention_heads, num_kv_heads, multiple_of, ffn_dim_multiplier, norm_eps, modulation=False, dtype=dtype, device=device, operations=operations ) for _ in range(num_refiner_layers) ]) self.layers = nn.ModuleList([ OmniGen2TransformerBlock( hidden_size, num_attention_heads, num_kv_heads, multiple_of, ffn_dim_multiplier, norm_eps, modulation=True, dtype=dtype, device=device, operations=operations ) for _ in range(num_layers) ]) self.norm_out = LuminaLayerNormContinuous( embedding_dim=hidden_size, conditioning_embedding_dim=min(hidden_size, 1024), elementwise_affine=False, eps=1e-6, out_dim=patch_size * patch_size * self.out_channels, dtype=dtype, device=device, operations=operations ) self.image_index_embedding = nn.Parameter(torch.empty(5, hidden_size, device=device, dtype=dtype)) def flat_and_pad_to_seq(self, hidden_states, ref_image_hidden_states): batch_size = len(hidden_states) p = self.patch_size img_sizes = [(img.size(1), img.size(2)) for img in hidden_states] l_effective_img_len = [(H // p) * (W // p) for (H, W) in img_sizes] if ref_image_hidden_states is not None: ref_image_hidden_states = list(map(lambda ref: comfy.ldm.common_dit.pad_to_patch_size(ref, (p, p)), ref_image_hidden_states)) ref_img_sizes = [[(imgs.size(2), imgs.size(3)) if imgs is not None else None for imgs in ref_image_hidden_states]] * batch_size l_effective_ref_img_len = [[(ref_img_size[0] // p) * (ref_img_size[1] // p) for ref_img_size in _ref_img_sizes] if _ref_img_sizes is not None else [0] for _ref_img_sizes in ref_img_sizes] else: ref_img_sizes = [None for _ in range(batch_size)] l_effective_ref_img_len = [[0] for _ in range(batch_size)] flat_ref_img_hidden_states = None if ref_image_hidden_states is not None: imgs = [] for ref_img in ref_image_hidden_states: B, C, H, W = ref_img.size() ref_img = rearrange(ref_img, 'b c (h p1) (w p2) -> b (h w) (p1 p2 c)', p1=p, p2=p) imgs.append(ref_img) flat_ref_img_hidden_states = torch.cat(imgs, dim=1) img = hidden_states B, C, H, W = img.size() flat_hidden_states = rearrange(img, 'b c (h p1) (w p2) -> b (h w) (p1 p2 c)', p1=p, p2=p) return ( flat_hidden_states, flat_ref_img_hidden_states, None, None, l_effective_ref_img_len, l_effective_img_len, ref_img_sizes, img_sizes, ) def img_patch_embed_and_refine(self, hidden_states, ref_image_hidden_states, padded_img_mask, padded_ref_img_mask, noise_rotary_emb, ref_img_rotary_emb, l_effective_ref_img_len, l_effective_img_len, temb, transformer_options={}): batch_size = len(hidden_states) hidden_states = self.x_embedder(hidden_states) if ref_image_hidden_states is not None: ref_image_hidden_states = self.ref_image_patch_embedder(ref_image_hidden_states) image_index_embedding = comfy.model_management.cast_to(self.image_index_embedding, dtype=hidden_states.dtype, device=hidden_states.device) for i in range(batch_size): shift = 0 for j, ref_img_len in enumerate(l_effective_ref_img_len[i]): ref_image_hidden_states[i, shift:shift + ref_img_len, :] = ref_image_hidden_states[i, shift:shift + ref_img_len, :] + image_index_embedding[j] shift += ref_img_len for layer in self.noise_refiner: hidden_states = layer(hidden_states, padded_img_mask, noise_rotary_emb, temb, transformer_options=transformer_options) if ref_image_hidden_states is not None: for layer in self.ref_image_refiner: ref_image_hidden_states = layer(ref_image_hidden_states, padded_ref_img_mask, ref_img_rotary_emb, temb, transformer_options=transformer_options) hidden_states = torch.cat([ref_image_hidden_states, hidden_states], dim=1) return hidden_states def forward(self, x, timesteps, context, num_tokens, ref_latents=None, attention_mask=None, transformer_options={}, **kwargs): B, C, H, W = x.shape hidden_states = comfy.ldm.common_dit.pad_to_patch_size(x, (self.patch_size, self.patch_size)) _, _, H_padded, W_padded = hidden_states.shape timestep = 1.0 - timesteps text_hidden_states = context text_attention_mask = attention_mask ref_image_hidden_states = ref_latents device = hidden_states.device temb, text_hidden_states = self.time_caption_embed(timestep, text_hidden_states, hidden_states[0].dtype) ( hidden_states, ref_image_hidden_states, img_mask, ref_img_mask, l_effective_ref_img_len, l_effective_img_len, ref_img_sizes, img_sizes, ) = self.flat_and_pad_to_seq(hidden_states, ref_image_hidden_states) ( context_rotary_emb, ref_img_rotary_emb, noise_rotary_emb, rotary_emb, encoder_seq_lengths, seq_lengths, ) = self.rope_embedder( hidden_states.shape[0], text_hidden_states.shape[1], [num_tokens] * text_hidden_states.shape[0], l_effective_ref_img_len, l_effective_img_len, ref_img_sizes, img_sizes, device, ) for layer in self.context_refiner: text_hidden_states = layer(text_hidden_states, text_attention_mask, context_rotary_emb, transformer_options=transformer_options) img_len = hidden_states.shape[1] combined_img_hidden_states = self.img_patch_embed_and_refine( hidden_states, ref_image_hidden_states, img_mask, ref_img_mask, noise_rotary_emb, ref_img_rotary_emb, l_effective_ref_img_len, l_effective_img_len, temb, transformer_options=transformer_options, ) hidden_states = torch.cat([text_hidden_states, combined_img_hidden_states], dim=1) attention_mask = None for layer in self.layers: hidden_states = layer(hidden_states, attention_mask, rotary_emb, temb, transformer_options=transformer_options) hidden_states = self.norm_out(hidden_states, temb) p = self.patch_size output = rearrange(hidden_states[:, -img_len:], 'b (h w) (p1 p2 c) -> b c (h p1) (w p2)', h=H_padded // p, w=W_padded// p, p1=p, p2=p)[:, :, :H, :W] return -output
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/ldm/omnigen/omnigen2.py", "license": "GNU General Public License v3.0", "lines": 372, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy/text_encoders/omnigen2.py
from transformers import Qwen2Tokenizer from comfy import sd1_clip import comfy.text_encoders.llama import os class Qwen25_3BTokenizer(sd1_clip.SDTokenizer): def __init__(self, embedding_directory=None, tokenizer_data={}): tokenizer_path = os.path.join(os.path.dirname(os.path.realpath(__file__)), "qwen25_tokenizer") super().__init__(tokenizer_path, pad_with_end=False, embedding_size=2048, embedding_key='qwen25_3b', tokenizer_class=Qwen2Tokenizer, has_start_token=False, has_end_token=False, pad_to_max_length=False, max_length=99999999, min_length=1, pad_token=151643, tokenizer_data=tokenizer_data) class Omnigen2Tokenizer(sd1_clip.SD1Tokenizer): def __init__(self, embedding_directory=None, tokenizer_data={}): super().__init__(embedding_directory=embedding_directory, tokenizer_data=tokenizer_data, name="qwen25_3b", tokenizer=Qwen25_3BTokenizer) self.llama_template = '<|im_start|>system\nYou are a helpful assistant that generates high-quality images based on user instructions.<|im_end|>\n<|im_start|>user\n{}<|im_end|>\n' def tokenize_with_weights(self, text, return_word_ids=False, llama_template=None,**kwargs): if llama_template is None: llama_text = self.llama_template.format(text) else: llama_text = llama_template.format(text) return super().tokenize_with_weights(llama_text, return_word_ids=return_word_ids, **kwargs) class Qwen25_3BModel(sd1_clip.SDClipModel): def __init__(self, device="cpu", layer="last", layer_idx=None, dtype=None, attention_mask=True, model_options={}): super().__init__(device=device, layer=layer, layer_idx=layer_idx, textmodel_json_config={}, dtype=dtype, special_tokens={"pad": 151643}, layer_norm_hidden_state=False, model_class=comfy.text_encoders.llama.Qwen25_3B, enable_attention_masks=attention_mask, return_attention_masks=attention_mask, model_options=model_options) class Omnigen2Model(sd1_clip.SD1ClipModel): def __init__(self, device="cpu", dtype=None, model_options={}): super().__init__(device=device, dtype=dtype, name="qwen25_3b", clip_model=Qwen25_3BModel, model_options=model_options) def te(dtype_llama=None, llama_quantization_metadata=None): class Omnigen2TEModel_(Omnigen2Model): def __init__(self, device="cpu", dtype=None, model_options={}): if llama_quantization_metadata is not None: model_options = model_options.copy() model_options["quantization_metadata"] = llama_quantization_metadata if dtype_llama is not None: dtype = dtype_llama super().__init__(device=device, dtype=dtype, model_options=model_options) return Omnigen2TEModel_
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/text_encoders/omnigen2.py", "license": "GNU General Public License v3.0", "lines": 34, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy_extras/nodes_edit_model.py
import node_helpers from typing_extensions import override from comfy_api.latest import ComfyExtension, io class ReferenceLatent(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="ReferenceLatent", category="advanced/conditioning/edit_models", description="This node sets the guiding latent for an edit model. If the model supports it you can chain multiple to set multiple reference images.", inputs=[ io.Conditioning.Input("conditioning"), io.Latent.Input("latent", optional=True), ], outputs=[ io.Conditioning.Output(), ] ) @classmethod def execute(cls, conditioning, latent=None) -> io.NodeOutput: if latent is not None: conditioning = node_helpers.conditioning_set_values(conditioning, {"reference_latents": [latent["samples"]]}, append=True) return io.NodeOutput(conditioning) class EditModelExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[io.ComfyNode]]: return [ ReferenceLatent, ] def comfy_entrypoint() -> EditModelExtension: return EditModelExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_extras/nodes_edit_model.py", "license": "GNU General Public License v3.0", "lines": 31, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy_extras/nodes_train.py
import logging import os import numpy as np import safetensors import torch import torch.nn as nn import torch.utils.checkpoint from tqdm.auto import trange from PIL import Image, ImageDraw, ImageFont from typing_extensions import override import comfy.samplers import comfy.sampler_helpers import comfy.sd import comfy.utils import comfy.model_management import comfy_extras.nodes_custom_sampler import folder_paths import node_helpers from comfy.weight_adapter import adapters, adapter_maps from comfy.weight_adapter.bypass import BypassInjectionManager from comfy_api.latest import ComfyExtension, io, ui from comfy.utils import ProgressBar class TrainGuider(comfy_extras.nodes_custom_sampler.Guider_Basic): """ CFGGuider with modifications for training specific logic """ def __init__(self, *args, offloading=False, **kwargs): super().__init__(*args, **kwargs) self.offloading = offloading def outer_sample( self, noise, latent_image, sampler, sigmas, denoise_mask=None, callback=None, disable_pbar=False, seed=None, latent_shapes=None, ): self.inner_model, self.conds, self.loaded_models = ( comfy.sampler_helpers.prepare_sampling( self.model_patcher, noise.shape, self.conds, self.model_options, force_full_load=not self.offloading, force_offload=self.offloading, ) ) torch.cuda.empty_cache() device = self.model_patcher.load_device if denoise_mask is not None: denoise_mask = comfy.sampler_helpers.prepare_mask( denoise_mask, noise.shape, device ) noise = noise.to(device) latent_image = latent_image.to(device) sigmas = sigmas.to(device) comfy.samplers.cast_to_load_options( self.model_options, device=device, dtype=self.model_patcher.model_dtype() ) try: self.model_patcher.pre_run() output = self.inner_sample( noise, latent_image, device, sampler, sigmas, denoise_mask, callback, disable_pbar, seed, latent_shapes=latent_shapes, ) finally: self.model_patcher.cleanup() comfy.sampler_helpers.cleanup_models(self.conds, self.loaded_models) del self.inner_model del self.loaded_models return output def make_batch_extra_option_dict(d, indicies, full_size=None): new_dict = {} for k, v in d.items(): newv = v if isinstance(v, dict): newv = make_batch_extra_option_dict(v, indicies, full_size=full_size) elif isinstance(v, torch.Tensor): if full_size is None or v.size(0) == full_size: newv = v[indicies] elif isinstance(v, (list, tuple)) and len(v) == full_size: newv = [v[i] for i in indicies] new_dict[k] = newv return new_dict def process_cond_list(d, prefix=""): if hasattr(d, "__iter__") and not hasattr(d, "items"): for index, item in enumerate(d): process_cond_list(item, f"{prefix}.{index}") return d elif hasattr(d, "items"): for k, v in list(d.items()): if isinstance(v, dict): process_cond_list(v, f"{prefix}.{k}") elif isinstance(v, torch.Tensor): d[k] = v.clone() elif isinstance(v, (list, tuple)): for index, item in enumerate(v): process_cond_list(item, f"{prefix}.{k}.{index}") return d class TrainSampler(comfy.samplers.Sampler): def __init__( self, loss_fn, optimizer, loss_callback=None, batch_size=1, grad_acc=1, total_steps=1, seed=0, training_dtype=torch.bfloat16, real_dataset=None, bucket_latents=None, ): self.loss_fn = loss_fn self.optimizer = optimizer self.loss_callback = loss_callback self.batch_size = batch_size self.total_steps = total_steps self.grad_acc = grad_acc self.seed = seed self.training_dtype = training_dtype self.real_dataset: list[torch.Tensor] | None = real_dataset # Bucket mode data self.bucket_latents: list[torch.Tensor] | None = ( bucket_latents # list of (Bi, C, Hi, Wi) ) # Precompute bucket offsets and weights for sampling if bucket_latents is not None: self._init_bucket_data(bucket_latents) else: self.bucket_offsets = None self.bucket_weights = None self.num_images = None def _init_bucket_data(self, bucket_latents): """Initialize bucket offsets and weights for sampling.""" self.bucket_offsets = [0] bucket_sizes = [] for lat in bucket_latents: bucket_sizes.append(lat.shape[0]) self.bucket_offsets.append(self.bucket_offsets[-1] + lat.shape[0]) self.num_images = self.bucket_offsets[-1] # Weights for sampling buckets proportional to their size self.bucket_weights = torch.tensor(bucket_sizes, dtype=torch.float32) def fwd_bwd( self, model_wrap, batch_sigmas, batch_noise, batch_latent, cond, indicies, extra_args, dataset_size, bwd=True, ): xt = model_wrap.inner_model.model_sampling.noise_scaling( batch_sigmas, batch_noise, batch_latent, False ) x0 = model_wrap.inner_model.model_sampling.noise_scaling( torch.zeros_like(batch_sigmas), torch.zeros_like(batch_noise), batch_latent, False, ) model_wrap.conds["positive"] = [cond[i] for i in indicies] batch_extra_args = make_batch_extra_option_dict( extra_args, indicies, full_size=dataset_size ) with torch.autocast(xt.device.type, dtype=self.training_dtype): x0_pred = model_wrap( xt.requires_grad_(True), batch_sigmas.requires_grad_(True), **batch_extra_args, ) loss = self.loss_fn(x0_pred, x0) if bwd: bwd_loss = loss / self.grad_acc bwd_loss.backward() return loss def _generate_batch_sigmas(self, model_wrap, batch_size, device): """Generate random sigma values for a batch.""" batch_sigmas = [ model_wrap.inner_model.model_sampling.percent_to_sigma( torch.rand((1,)).item() ) for _ in range(batch_size) ] return torch.tensor(batch_sigmas).to(device) def _train_step_bucket_mode(self, model_wrap, cond, extra_args, noisegen, latent_image, pbar): """Execute one training step in bucket mode.""" # Sample bucket (weighted by size), then sample batch from bucket bucket_idx = torch.multinomial(self.bucket_weights, 1).item() bucket_latent = self.bucket_latents[bucket_idx] # (Bi, C, Hi, Wi) bucket_size = bucket_latent.shape[0] bucket_offset = self.bucket_offsets[bucket_idx] # Sample indices from this bucket (use all if bucket_size < batch_size) actual_batch_size = min(self.batch_size, bucket_size) relative_indices = torch.randperm(bucket_size)[:actual_batch_size].tolist() # Convert to absolute indices for fwd_bwd (cond is flattened, use absolute index) absolute_indices = [bucket_offset + idx for idx in relative_indices] batch_latent = bucket_latent[relative_indices].to(latent_image) # (actual_batch_size, C, H, W) batch_noise = noisegen.generate_noise({"samples": batch_latent}).to( batch_latent.device ) batch_sigmas = self._generate_batch_sigmas(model_wrap, actual_batch_size, batch_latent.device) loss = self.fwd_bwd( model_wrap, batch_sigmas, batch_noise, batch_latent, cond, # Use flattened cond with absolute indices absolute_indices, extra_args, self.num_images, bwd=True, ) if self.loss_callback: self.loss_callback(loss.item()) pbar.set_postfix({"loss": f"{loss.item():.4f}", "bucket": bucket_idx}) def _train_step_standard_mode(self, model_wrap, cond, extra_args, noisegen, latent_image, dataset_size, pbar): """Execute one training step in standard (non-bucket, non-multi-res) mode.""" indicies = torch.randperm(dataset_size)[: self.batch_size].tolist() batch_latent = torch.stack([latent_image[i] for i in indicies]) batch_noise = noisegen.generate_noise({"samples": batch_latent}).to( batch_latent.device ) batch_sigmas = self._generate_batch_sigmas(model_wrap, min(self.batch_size, dataset_size), batch_latent.device) loss = self.fwd_bwd( model_wrap, batch_sigmas, batch_noise, batch_latent, cond, indicies, extra_args, dataset_size, bwd=True, ) if self.loss_callback: self.loss_callback(loss.item()) pbar.set_postfix({"loss": f"{loss.item():.4f}"}) def _train_step_multires_mode(self, model_wrap, cond, extra_args, noisegen, latent_image, dataset_size, pbar): """Execute one training step in multi-resolution mode (real_dataset is set).""" indicies = torch.randperm(dataset_size)[: self.batch_size].tolist() total_loss = 0 for index in indicies: single_latent = self.real_dataset[index].to(latent_image) batch_noise = noisegen.generate_noise( {"samples": single_latent} ).to(single_latent.device) batch_sigmas = ( model_wrap.inner_model.model_sampling.percent_to_sigma( torch.rand((1,)).item() ) ) batch_sigmas = torch.tensor([batch_sigmas]).to(single_latent.device) loss = self.fwd_bwd( model_wrap, batch_sigmas, batch_noise, single_latent, cond, [index], extra_args, dataset_size, bwd=False, ) total_loss += loss total_loss = total_loss / self.grad_acc / len(indicies) total_loss.backward() if self.loss_callback: self.loss_callback(total_loss.item()) pbar.set_postfix({"loss": f"{total_loss.item():.4f}"}) def sample( self, model_wrap, sigmas, extra_args, callback, noise, latent_image=None, denoise_mask=None, disable_pbar=False, ): model_wrap.conds = process_cond_list(model_wrap.conds) cond = model_wrap.conds["positive"] dataset_size = sigmas.size(0) torch.cuda.empty_cache() ui_pbar = ProgressBar(self.total_steps) for i in ( pbar := trange( self.total_steps, desc="Training LoRA", smoothing=0.01, disable=not comfy.utils.PROGRESS_BAR_ENABLED, ) ): noisegen = comfy_extras.nodes_custom_sampler.Noise_RandomNoise( self.seed + i * 1000 ) if self.bucket_latents is not None: self._train_step_bucket_mode(model_wrap, cond, extra_args, noisegen, latent_image, pbar) elif self.real_dataset is None: self._train_step_standard_mode(model_wrap, cond, extra_args, noisegen, latent_image, dataset_size, pbar) else: self._train_step_multires_mode(model_wrap, cond, extra_args, noisegen, latent_image, dataset_size, pbar) if (i + 1) % self.grad_acc == 0: for param_groups in self.optimizer.param_groups: for param in param_groups["params"]: if param.grad is None: continue param.grad.data = param.grad.data.to(param.data.dtype) self.optimizer.step() self.optimizer.zero_grad() ui_pbar.update(1) torch.cuda.empty_cache() return torch.zeros_like(latent_image) class BiasDiff(torch.nn.Module): def __init__(self, bias): super().__init__() self.bias = bias def __call__(self, b): org_dtype = b.dtype return (b.to(self.bias) + self.bias).to(org_dtype) def passive_memory_usage(self): return self.bias.nelement() * self.bias.element_size() def move_to(self, device): self.to(device=device) return self.passive_memory_usage() def draw_loss_graph(loss_map, steps): width, height = 500, 300 img = Image.new("RGB", (width, height), "white") draw = ImageDraw.Draw(img) min_loss, max_loss = min(loss_map.values()), max(loss_map.values()) scaled_loss = [(l - min_loss) / (max_loss - min_loss) for l in loss_map.values()] prev_point = (0, height - int(scaled_loss[0] * height)) for i, l in enumerate(scaled_loss[1:], start=1): x = int(i / (steps - 1) * width) y = height - int(l * height) draw.line([prev_point, (x, y)], fill="blue", width=2) prev_point = (x, y) return img def find_all_highest_child_module_with_forward( model: torch.nn.Module, result=None, name=None ): if result is None: result = [] elif hasattr(model, "forward") and not isinstance( model, (torch.nn.ModuleList, torch.nn.Sequential, torch.nn.ModuleDict) ): result.append(model) logging.debug(f"Found module with forward: {name} ({model.__class__.__name__})") return result name = name or "root" for next_name, child in model.named_children(): find_all_highest_child_module_with_forward(child, result, f"{name}.{next_name}") return result def find_modules_at_depth( model: nn.Module, depth: int = 1, result=None, current_depth=0, name=None ) -> list[nn.Module]: """ Find modules at a specific depth level for gradient checkpointing. Args: model: The model to search depth: Target depth level (1 = top-level blocks, 2 = their children, etc.) result: Accumulator for results current_depth: Current recursion depth name: Current module name for logging Returns: List of modules at the target depth """ if result is None: result = [] name = name or "root" # Skip container modules (they don't have meaningful forward) is_container = isinstance(model, (nn.ModuleList, nn.Sequential, nn.ModuleDict)) has_forward = hasattr(model, "forward") and not is_container if has_forward: current_depth += 1 if current_depth == depth: result.append(model) logging.debug(f"Found module at depth {depth}: {name} ({model.__class__.__name__})") return result # Recurse into children for next_name, child in model.named_children(): find_modules_at_depth(child, depth, result, current_depth, f"{name}.{next_name}") return result class OffloadCheckpointFunction(torch.autograd.Function): """ Gradient checkpointing that works with weight offloading. Forward: no_grad -> compute -> weights can be freed Backward: enable_grad -> recompute -> backward -> weights can be freed For single input, single output modules (Linear, Conv*). """ @staticmethod def forward(ctx, x: torch.Tensor, forward_fn): ctx.save_for_backward(x) ctx.forward_fn = forward_fn with torch.no_grad(): return forward_fn(x) @staticmethod def backward(ctx, grad_out: torch.Tensor): x, = ctx.saved_tensors forward_fn = ctx.forward_fn # Clear context early ctx.forward_fn = None with torch.enable_grad(): x_detached = x.detach().requires_grad_(True) y = forward_fn(x_detached) y.backward(grad_out) grad_x = x_detached.grad # Explicit cleanup del y, x_detached, forward_fn return grad_x, None def patch(m, offloading=False): if not hasattr(m, "forward"): return org_forward = m.forward # Branch 1: Linear/Conv* -> offload-compatible checkpoint (single input/output) if offloading and isinstance(m, (nn.Linear, nn.Conv1d, nn.Conv2d, nn.Conv3d)): def checkpointing_fwd(x): return OffloadCheckpointFunction.apply(x, org_forward) # Branch 2: Others -> standard checkpoint else: def fwd(args, kwargs): return org_forward(*args, **kwargs) def checkpointing_fwd(*args, **kwargs): return torch.utils.checkpoint.checkpoint(fwd, args, kwargs, use_reentrant=False) m.org_forward = org_forward m.forward = checkpointing_fwd def unpatch(m): if hasattr(m, "org_forward"): m.forward = m.org_forward del m.org_forward def _process_latents_bucket_mode(latents): """Process latents for bucket mode training. Args: latents: list[{"samples": tensor}] where each tensor is (Bi, C, Hi, Wi) Returns: list of latent tensors """ bucket_latents = [] for latent_dict in latents: bucket_latents.append(latent_dict["samples"]) # (Bi, C, Hi, Wi) return bucket_latents def _process_latents_standard_mode(latents): """Process latents for standard (non-bucket) mode training. Args: latents: list of latent dicts or single latent dict Returns: Processed latents (tensor or list of tensors) """ if len(latents) == 1: return latents[0]["samples"] # Single latent dict latent_list = [] for latent in latents: latent = latent["samples"] bs = latent.shape[0] if bs != 1: for sub_latent in latent: latent_list.append(sub_latent[None]) else: latent_list.append(latent) return latent_list def _process_conditioning(positive): """Process conditioning - either single list or list of lists. Args: positive: list of conditioning Returns: Flattened conditioning list """ if len(positive) == 1: return positive[0] # Single conditioning list # Multiple conditioning lists - flatten flat_positive = [] for cond in positive: if isinstance(cond, list): flat_positive.extend(cond) else: flat_positive.append(cond) return flat_positive def _prepare_latents_and_count(latents, dtype, bucket_mode): """Convert latents to dtype and compute image counts. Args: latents: Latents (tensor, list of tensors, or bucket list) dtype: Target dtype bucket_mode: Whether bucket mode is enabled Returns: tuple: (processed_latents, num_images, multi_res) """ if bucket_mode: # In bucket mode, latents is list of tensors (Bi, C, Hi, Wi) latents = [t.to(dtype) for t in latents] num_buckets = len(latents) num_images = sum(t.shape[0] for t in latents) multi_res = False # Not using multi_res path in bucket mode logging.debug(f"Bucket mode: {num_buckets} buckets, {num_images} total samples") for i, lat in enumerate(latents): logging.debug(f" Bucket {i}: shape {lat.shape}") return latents, num_images, multi_res # Non-bucket mode if isinstance(latents, list): all_shapes = set() latents = [t.to(dtype) for t in latents] for latent in latents: all_shapes.add(latent.shape) logging.debug(f"Latent shapes: {all_shapes}") if len(all_shapes) > 1: multi_res = True else: multi_res = False latents = torch.cat(latents, dim=0) num_images = len(latents) elif isinstance(latents, torch.Tensor): latents = latents.to(dtype) num_images = latents.shape[0] multi_res = False else: logging.error(f"Invalid latents type: {type(latents)}") num_images = 0 multi_res = False return latents, num_images, multi_res def _validate_and_expand_conditioning(positive, num_images, bucket_mode): """Validate conditioning count matches image count, expand if needed. Args: positive: Conditioning list num_images: Number of images bucket_mode: Whether bucket mode is enabled Returns: Validated/expanded conditioning list Raises: ValueError: If conditioning count doesn't match image count """ if bucket_mode: return positive # Skip validation in bucket mode logging.debug(f"Total Images: {num_images}, Total Captions: {len(positive)}") if len(positive) == 1 and num_images > 1: return positive * num_images elif len(positive) != num_images: raise ValueError( f"Number of positive conditions ({len(positive)}) does not match number of images ({num_images})." ) return positive def _load_existing_lora(existing_lora): """Load existing LoRA weights if provided. Args: existing_lora: LoRA filename or "[None]" Returns: tuple: (existing_weights dict, existing_steps int) """ if existing_lora == "[None]": return {}, 0 lora_path = folder_paths.get_full_path_or_raise("loras", existing_lora) # Extract steps from filename like "trained_lora_10_steps_20250225_203716" existing_steps = int(existing_lora.split("_steps_")[0].split("_")[-1]) existing_weights = {} if lora_path: existing_weights = comfy.utils.load_torch_file(lora_path) return existing_weights, existing_steps def _create_weight_adapter( module, module_name, existing_weights, algorithm, lora_dtype, rank ): """Create a weight adapter for a module with weight. Args: module: The module to create adapter for module_name: Name of the module existing_weights: Dict of existing LoRA weights algorithm: Algorithm name for new adapters lora_dtype: dtype for LoRA weights rank: Rank for new LoRA adapters Returns: tuple: (train_adapter, lora_params dict) """ key = f"{module_name}.weight" shape = module.weight.shape lora_params = {} logging.debug(f"Creating weight adapter for {key} with shape {shape}") if len(shape) >= 2: alpha = float(existing_weights.get(f"{key}.alpha", 1.0)) dora_scale = existing_weights.get(f"{key}.dora_scale", None) # Try to load existing adapter existing_adapter = None for adapter_cls in adapters: existing_adapter = adapter_cls.load( module_name, existing_weights, alpha, dora_scale ) if existing_adapter is not None: break if existing_adapter is None: adapter_cls = adapter_maps[algorithm] if existing_adapter is not None: train_adapter = existing_adapter.to_train().to(lora_dtype) else: # Use LoRA with alpha=1.0 by default train_adapter = adapter_cls.create_train( module.weight, rank=rank, alpha=1.0 ).to(lora_dtype) for name, parameter in train_adapter.named_parameters(): lora_params[f"{module_name}.{name}"] = parameter return train_adapter.train().requires_grad_(True), lora_params else: # 1D weight - use BiasDiff diff = torch.nn.Parameter( torch.zeros(module.weight.shape, dtype=lora_dtype, requires_grad=True) ) diff_module = BiasDiff(diff).train().requires_grad_(True) lora_params[f"{module_name}.diff"] = diff return diff_module, lora_params def _create_bias_adapter(module, module_name, lora_dtype): """Create a bias adapter for a module with bias. Args: module: The module with bias module_name: Name of the module lora_dtype: dtype for LoRA weights Returns: tuple: (bias_module, lora_params dict) """ bias = torch.nn.Parameter( torch.zeros(module.bias.shape, dtype=lora_dtype, requires_grad=True) ) bias_module = BiasDiff(bias).train().requires_grad_(True) lora_params = {f"{module_name}.diff_b": bias} return bias_module, lora_params def _setup_lora_adapters(mp, existing_weights, algorithm, lora_dtype, rank): """Setup all LoRA adapters on the model. Args: mp: Model patcher existing_weights: Dict of existing LoRA weights algorithm: Algorithm name for new adapters lora_dtype: dtype for LoRA weights rank: Rank for new LoRA adapters Returns: tuple: (lora_sd dict, all_weight_adapters list) """ lora_sd = {} all_weight_adapters = [] for n, m in mp.model.named_modules(): if hasattr(m, "weight_function"): if m.weight is not None: adapter, params = _create_weight_adapter( m, n, existing_weights, algorithm, lora_dtype, rank ) lora_sd.update(params) key = f"{n}.weight" mp.add_weight_wrapper(key, adapter) all_weight_adapters.append(adapter) if hasattr(m, "bias") and m.bias is not None: bias_adapter, bias_params = _create_bias_adapter(m, n, lora_dtype) lora_sd.update(bias_params) key = f"{n}.bias" mp.add_weight_wrapper(key, bias_adapter) all_weight_adapters.append(bias_adapter) return lora_sd, all_weight_adapters def _setup_lora_adapters_bypass(mp, existing_weights, algorithm, lora_dtype, rank): """Setup LoRA adapters in bypass mode. In bypass mode: - Weight adapters (lora/lokr/oft) use bypass injection (forward hook) - Bias/norm adapters (BiasDiff) still use weight wrapper (direct modification) This is useful when the base model weights are quantized and cannot be directly modified. Args: mp: Model patcher existing_weights: Dict of existing LoRA weights algorithm: Algorithm name for new adapters lora_dtype: dtype for LoRA weights rank: Rank for new LoRA adapters Returns: tuple: (lora_sd dict, all_weight_adapters list, bypass_manager) """ lora_sd = {} all_weight_adapters = [] bypass_manager = BypassInjectionManager() for n, m in mp.model.named_modules(): if hasattr(m, "weight_function"): if m.weight is not None: adapter, params = _create_weight_adapter( m, n, existing_weights, algorithm, lora_dtype, rank ) lora_sd.update(params) all_weight_adapters.append(adapter) key = f"{n}.weight" # BiasDiff (for 1D weights like norm) uses weight wrapper, not bypass # Only use bypass for adapters that have h() method (lora/lokr/oft) if isinstance(adapter, BiasDiff): mp.add_weight_wrapper(key, adapter) logging.debug(f"[BypassMode] Added 1D weight adapter (weight wrapper) for {key}") else: bypass_manager.add_adapter(key, adapter, strength=1.0) logging.debug(f"[BypassMode] Added weight adapter (bypass) for {key}") if hasattr(m, "bias") and m.bias is not None: # Bias adapters still use weight wrapper (bias is usually not quantized) bias_adapter, bias_params = _create_bias_adapter(m, n, lora_dtype) lora_sd.update(bias_params) key = f"{n}.bias" mp.add_weight_wrapper(key, bias_adapter) all_weight_adapters.append(bias_adapter) logging.debug(f"[BypassMode] Added bias adapter (weight wrapper) for {key}") return lora_sd, all_weight_adapters, bypass_manager def _create_optimizer(optimizer_name, parameters, learning_rate): """Create optimizer based on name. Args: optimizer_name: Name of optimizer ("Adam", "AdamW", "SGD", "RMSprop") parameters: Parameters to optimize learning_rate: Learning rate Returns: Optimizer instance """ if optimizer_name == "Adam": return torch.optim.Adam(parameters, lr=learning_rate) elif optimizer_name == "AdamW": return torch.optim.AdamW(parameters, lr=learning_rate) elif optimizer_name == "SGD": return torch.optim.SGD(parameters, lr=learning_rate) elif optimizer_name == "RMSprop": return torch.optim.RMSprop(parameters, lr=learning_rate) def _create_loss_function(loss_function_name): """Create loss function based on name. Args: loss_function_name: Name of loss function ("MSE", "L1", "Huber", "SmoothL1") Returns: Loss function instance """ if loss_function_name == "MSE": return torch.nn.MSELoss() elif loss_function_name == "L1": return torch.nn.L1Loss() elif loss_function_name == "Huber": return torch.nn.HuberLoss() elif loss_function_name == "SmoothL1": return torch.nn.SmoothL1Loss() def _run_training_loop( guider, train_sampler, latents, num_images, seed, bucket_mode, multi_res ): """Execute the training loop. Args: guider: The guider object train_sampler: The training sampler latents: Latent tensors num_images: Number of images seed: Random seed bucket_mode: Whether bucket mode is enabled multi_res: Whether multi-resolution mode is enabled """ sigmas = torch.tensor(range(num_images)) noise = comfy_extras.nodes_custom_sampler.Noise_RandomNoise(seed) if bucket_mode: # Use first bucket's first latent as dummy for guider dummy_latent = latents[0][:1].repeat(num_images, 1, 1, 1) guider.sample( noise.generate_noise({"samples": dummy_latent}), dummy_latent, train_sampler, sigmas, seed=noise.seed, ) elif multi_res: # use first latent as dummy latent if multi_res latents = latents[0].repeat(num_images, 1, 1, 1) guider.sample( noise.generate_noise({"samples": latents}), latents, train_sampler, sigmas, seed=noise.seed, ) else: guider.sample( noise.generate_noise({"samples": latents}), latents, train_sampler, sigmas, seed=noise.seed, ) class TrainLoraNode(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="TrainLoraNode", display_name="Train LoRA", category="training", is_experimental=True, is_input_list=True, # All inputs become lists inputs=[ io.Model.Input("model", tooltip="The model to train the LoRA on."), io.Latent.Input( "latents", tooltip="The Latents to use for training, serve as dataset/input of the model.", ), io.Conditioning.Input( "positive", tooltip="The positive conditioning to use for training." ), io.Int.Input( "batch_size", default=1, min=1, max=10000, tooltip="The batch size to use for training.", ), io.Int.Input( "grad_accumulation_steps", default=1, min=1, max=1024, tooltip="The number of gradient accumulation steps to use for training.", ), io.Int.Input( "steps", default=16, min=1, max=100000, tooltip="The number of steps to train the LoRA for.", ), io.Float.Input( "learning_rate", default=0.0005, min=0.0000001, max=1.0, step=0.0000001, tooltip="The learning rate to use for training.", ), io.Int.Input( "rank", default=8, min=1, max=128, tooltip="The rank of the LoRA layers.", ), io.Combo.Input( "optimizer", options=["AdamW", "Adam", "SGD", "RMSprop"], default="AdamW", tooltip="The optimizer to use for training.", ), io.Combo.Input( "loss_function", options=["MSE", "L1", "Huber", "SmoothL1"], default="MSE", tooltip="The loss function to use for training.", ), io.Int.Input( "seed", default=0, min=0, max=0xFFFFFFFFFFFFFFFF, tooltip="The seed to use for training (used in generator for LoRA weight initialization and noise sampling)", ), io.Combo.Input( "training_dtype", options=["bf16", "fp32"], default="bf16", tooltip="The dtype to use for training.", ), io.Combo.Input( "lora_dtype", options=["bf16", "fp32"], default="bf16", tooltip="The dtype to use for lora.", ), io.Combo.Input( "algorithm", options=list(adapter_maps.keys()), default=list(adapter_maps.keys())[0], tooltip="The algorithm to use for training.", ), io.Boolean.Input( "gradient_checkpointing", default=True, tooltip="Use gradient checkpointing for training.", ), io.Int.Input( "checkpoint_depth", default=1, min=1, max=5, tooltip="Depth level for gradient checkpointing.", ), io.Boolean.Input( "offloading", default=False, tooltip="Offload the Model to RAM. Requires Bypass Mode.", ), io.Combo.Input( "existing_lora", options=folder_paths.get_filename_list("loras") + ["[None]"], default="[None]", tooltip="The existing LoRA to append to. Set to None for new LoRA.", ), io.Boolean.Input( "bucket_mode", default=False, tooltip="Enable resolution bucket mode. When enabled, expects pre-bucketed latents from ResolutionBucket node.", ), io.Boolean.Input( "bypass_mode", default=False, tooltip="Enable bypass mode for training. When enabled, adapters are applied via forward hooks instead of weight modification. Useful for quantized models where weights cannot be directly modified.", ), ], outputs=[ io.Custom("LORA_MODEL").Output( display_name="lora", tooltip="LoRA weights" ), io.Custom("LOSS_MAP").Output( display_name="loss_map", tooltip="Loss history" ), io.Int.Output(display_name="steps", tooltip="Total training steps"), ], ) @classmethod def execute( cls, model, latents, positive, batch_size, steps, grad_accumulation_steps, learning_rate, rank, optimizer, loss_function, seed, training_dtype, lora_dtype, algorithm, gradient_checkpointing, checkpoint_depth, offloading, existing_lora, bucket_mode, bypass_mode, ): # Extract scalars from lists (due to is_input_list=True) model = model[0] batch_size = batch_size[0] steps = steps[0] grad_accumulation_steps = grad_accumulation_steps[0] learning_rate = learning_rate[0] rank = rank[0] optimizer_name = optimizer[0] loss_function_name = loss_function[0] seed = seed[0] training_dtype = training_dtype[0] lora_dtype = lora_dtype[0] algorithm = algorithm[0] gradient_checkpointing = gradient_checkpointing[0] offloading = offloading[0] checkpoint_depth = checkpoint_depth[0] existing_lora = existing_lora[0] bucket_mode = bucket_mode[0] bypass_mode = bypass_mode[0] # Process latents based on mode if bucket_mode: latents = _process_latents_bucket_mode(latents) else: latents = _process_latents_standard_mode(latents) # Process conditioning positive = _process_conditioning(positive) # Setup model and dtype mp = model.clone() dtype = node_helpers.string_to_torch_dtype(training_dtype) lora_dtype = node_helpers.string_to_torch_dtype(lora_dtype) mp.set_model_compute_dtype(dtype) if mp.is_dynamic(): if not bypass_mode: logging.info("Training MP is Dynamic - forcing bypass mode. Start comfy with --highvram to force weight diff mode") bypass_mode = True offloading = True elif offloading: if not bypass_mode: logging.info("Training Offload selected - forcing bypass mode. Set bypass = True to remove this message") # Prepare latents and compute counts latents, num_images, multi_res = _prepare_latents_and_count( latents, dtype, bucket_mode ) # Validate and expand conditioning positive = _validate_and_expand_conditioning(positive, num_images, bucket_mode) with torch.inference_mode(False): # Setup models for training mp.model.requires_grad_(False) # Load existing LoRA weights if provided existing_weights, existing_steps = _load_existing_lora(existing_lora) # Setup LoRA adapters bypass_manager = None if bypass_mode: logging.debug("Using bypass mode for training") lora_sd, all_weight_adapters, bypass_manager = _setup_lora_adapters_bypass( mp, existing_weights, algorithm, lora_dtype, rank ) else: lora_sd, all_weight_adapters = _setup_lora_adapters( mp, existing_weights, algorithm, lora_dtype, rank ) # Create optimizer and loss function optimizer = _create_optimizer( optimizer_name, lora_sd.values(), learning_rate ) criterion = _create_loss_function(loss_function_name) # Setup gradient checkpointing if gradient_checkpointing: modules_to_patch = find_modules_at_depth( mp.model.diffusion_model, depth=checkpoint_depth ) logging.info(f"Gradient checkpointing: patching {len(modules_to_patch)} modules at depth {checkpoint_depth}") for m in modules_to_patch: patch(m, offloading=offloading) torch.cuda.empty_cache() # With force_full_load=False we should be able to have offloading # But for offloading in training we need custom AutoGrad hooks for fwd/bwd comfy.model_management.load_models_gpu( [mp], memory_required=1e20, force_full_load=not offloading ) torch.cuda.empty_cache() # Setup loss tracking loss_map = {"loss": []} def loss_callback(loss): loss_map["loss"].append(loss) # Create sampler if bucket_mode: train_sampler = TrainSampler( criterion, optimizer, loss_callback=loss_callback, batch_size=batch_size, grad_acc=grad_accumulation_steps, total_steps=steps * grad_accumulation_steps, seed=seed, training_dtype=dtype, bucket_latents=latents, ) else: train_sampler = TrainSampler( criterion, optimizer, loss_callback=loss_callback, batch_size=batch_size, grad_acc=grad_accumulation_steps, total_steps=steps * grad_accumulation_steps, seed=seed, training_dtype=dtype, real_dataset=latents if multi_res else None, ) # Setup guider guider = TrainGuider(mp, offloading=offloading) guider.set_conds(positive) # Inject bypass hooks if bypass mode is enabled bypass_injections = None if bypass_manager is not None: bypass_injections = bypass_manager.create_injections(mp.model) for injection in bypass_injections: injection.inject(mp) logging.debug(f"[BypassMode] Injected {bypass_manager.get_hook_count()} bypass hooks") # Run training loop try: comfy.model_management.in_training = True _run_training_loop( guider, train_sampler, latents, num_images, seed, bucket_mode, multi_res, ) finally: comfy.model_management.in_training = False # Eject bypass hooks if they were injected if bypass_injections is not None: for injection in bypass_injections: injection.eject(mp) logging.debug("[BypassMode] Ejected bypass hooks") for m in mp.model.modules(): unpatch(m) del train_sampler, optimizer for param in lora_sd: lora_sd[param] = lora_sd[param].to(lora_dtype).detach() for adapter in all_weight_adapters: adapter.requires_grad_(False) del adapter del all_weight_adapters # mp in train node is highly specialized for training # use it in inference will result in bad behavior so we don't return it return io.NodeOutput(lora_sd, loss_map, steps + existing_steps) class LoraModelLoader(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="LoraModelLoader", display_name="Load LoRA Model", category="loaders", is_experimental=True, inputs=[ io.Model.Input( "model", tooltip="The diffusion model the LoRA will be applied to." ), io.Custom("LORA_MODEL").Input( "lora", tooltip="The LoRA model to apply to the diffusion model." ), io.Float.Input( "strength_model", default=1.0, min=-100.0, max=100.0, tooltip="How strongly to modify the diffusion model. This value can be negative.", ), io.Boolean.Input( "bypass", default=False, tooltip="When enabled, applies LoRA in bypass mode without modifying base model weights. Useful for training and when model weights are offloaded.", ), ], outputs=[ io.Model.Output( display_name="model", tooltip="The modified diffusion model." ), ], ) @classmethod def execute(cls, model, lora, strength_model, bypass=False): if strength_model == 0: return io.NodeOutput(model) if bypass: model_lora, _ = comfy.sd.load_bypass_lora_for_models( model, None, lora, strength_model, 0 ) else: model_lora, _ = comfy.sd.load_lora_for_models( model, None, lora, strength_model, 0 ) return io.NodeOutput(model_lora) class SaveLoRA(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="SaveLoRA", search_aliases=["export lora"], display_name="Save LoRA Weights", category="loaders", is_experimental=True, is_output_node=True, inputs=[ io.Custom("LORA_MODEL").Input( "lora", tooltip="The LoRA model to save. Do not use the model with LoRA layers.", ), io.String.Input( "prefix", default="loras/ComfyUI_trained_lora", tooltip="The prefix to use for the saved LoRA file.", ), io.Int.Input( "steps", optional=True, tooltip="Optional: The number of steps to LoRA has been trained for, used to name the saved file.", ), ], outputs=[], ) @classmethod def execute(cls, lora, prefix, steps=None): output_dir = folder_paths.get_output_directory() full_output_folder, filename, counter, subfolder, filename_prefix = ( folder_paths.get_save_image_path(prefix, output_dir) ) if steps is None: output_checkpoint = f"{filename}_{counter:05}_.safetensors" else: output_checkpoint = f"{filename}_{steps}_steps_{counter:05}_.safetensors" output_checkpoint = os.path.join(full_output_folder, output_checkpoint) safetensors.torch.save_file(lora, output_checkpoint) return io.NodeOutput() class LossGraphNode(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="LossGraphNode", search_aliases=["training chart", "training visualization", "plot loss"], display_name="Plot Loss Graph", category="training", is_experimental=True, is_output_node=True, inputs=[ io.Custom("LOSS_MAP").Input( "loss", tooltip="Loss map from training node." ), io.String.Input( "filename_prefix", default="loss_graph", tooltip="Prefix for the saved loss graph image.", ), ], outputs=[], hidden=[io.Hidden.prompt, io.Hidden.extra_pnginfo], ) @classmethod def execute(cls, loss, filename_prefix, prompt=None, extra_pnginfo=None): loss_values = loss["loss"] width, height = 800, 480 margin = 40 img = Image.new( "RGB", (width + margin, height + margin), "white" ) # Extend canvas draw = ImageDraw.Draw(img) min_loss, max_loss = min(loss_values), max(loss_values) scaled_loss = [(l - min_loss) / (max_loss - min_loss) for l in loss_values] steps = len(loss_values) prev_point = (margin, height - int(scaled_loss[0] * height)) for i, l in enumerate(scaled_loss[1:], start=1): x = margin + int(i / steps * width) # Scale X properly y = height - int(l * height) draw.line([prev_point, (x, y)], fill="blue", width=2) prev_point = (x, y) draw.line([(margin, 0), (margin, height)], fill="black", width=2) # Y-axis draw.line( [(margin, height), (width + margin, height)], fill="black", width=2 ) # X-axis font = None try: font = ImageFont.truetype("arial.ttf", 12) except IOError: font = ImageFont.load_default() # Add axis labels draw.text((5, height // 2), "Loss", font=font, fill="black") draw.text((width // 2, height + 10), "Steps", font=font, fill="black") # Add min/max loss values draw.text((margin - 30, 0), f"{max_loss:.2f}", font=font, fill="black") draw.text( (margin - 30, height - 10), f"{min_loss:.2f}", font=font, fill="black" ) # Convert PIL image to tensor for PreviewImage img_array = np.array(img).astype(np.float32) / 255.0 img_tensor = torch.from_numpy(img_array)[None,] # [1, H, W, 3] # Return preview UI return io.NodeOutput(ui=ui.PreviewImage(img_tensor, cls=cls)) # ========== Extension Setup ========== class TrainingExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[io.ComfyNode]]: return [ TrainLoraNode, LoraModelLoader, SaveLoRA, LossGraphNode, ] async def comfy_entrypoint() -> TrainingExtension: return TrainingExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_extras/nodes_train.py", "license": "GNU General Public License v3.0", "lines": 1258, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:tests-unit/folder_paths_test/misc_test.py
import pytest import os import tempfile from folder_paths import get_input_subfolders, set_input_directory @pytest.fixture(scope="module") def mock_folder_structure(): with tempfile.TemporaryDirectory() as temp_dir: # Create a nested folder structure folders = [ "folder1", "folder1/subfolder1", "folder1/subfolder2", "folder2", "folder2/deep", "folder2/deep/nested", "empty_folder" ] # Create the folders for folder in folders: os.makedirs(os.path.join(temp_dir, folder)) # Add some files to test they're not included with open(os.path.join(temp_dir, "root_file.txt"), "w") as f: f.write("test") with open(os.path.join(temp_dir, "folder1", "test.txt"), "w") as f: f.write("test") set_input_directory(temp_dir) yield temp_dir def test_gets_all_folders(mock_folder_structure): folders = get_input_subfolders() expected = ["folder1", "folder1/subfolder1", "folder1/subfolder2", "folder2", "folder2/deep", "folder2/deep/nested", "empty_folder"] assert sorted(folders) == sorted(expected) def test_handles_nonexistent_input_directory(): with tempfile.TemporaryDirectory() as temp_dir: nonexistent = os.path.join(temp_dir, "nonexistent") set_input_directory(nonexistent) assert get_input_subfolders() == [] def test_empty_input_directory(): with tempfile.TemporaryDirectory() as temp_dir: set_input_directory(temp_dir) assert get_input_subfolders() == [] # Empty since we don't include root
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "tests-unit/folder_paths_test/misc_test.py", "license": "GNU General Public License v3.0", "lines": 41, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
Comfy-Org/ComfyUI:comfy/ldm/cosmos/predict2.py
# original code from: https://github.com/nvidia-cosmos/cosmos-predict2 import torch from torch import nn from einops import rearrange from einops.layers.torch import Rearrange import logging from typing import Callable, Optional, Tuple import math from .position_embedding import VideoRopePosition3DEmb, LearnablePosEmbAxis from torchvision import transforms import comfy.patcher_extension from comfy.ldm.modules.attention import optimized_attention import comfy.ldm.common_dit def apply_rotary_pos_emb( t: torch.Tensor, freqs: torch.Tensor, ) -> torch.Tensor: t_ = t.reshape(*t.shape[:-1], 2, -1).movedim(-2, -1).unsqueeze(-2).float() t_out = freqs[..., 0] * t_[..., 0] + freqs[..., 1] * t_[..., 1] t_out = t_out.movedim(-1, -2).reshape(*t.shape).type_as(t) return t_out # ---------------------- Feed Forward Network ----------------------- class GPT2FeedForward(nn.Module): def __init__(self, d_model: int, d_ff: int, device=None, dtype=None, operations=None) -> None: super().__init__() self.activation = nn.GELU() self.layer1 = operations.Linear(d_model, d_ff, bias=False, device=device, dtype=dtype) self.layer2 = operations.Linear(d_ff, d_model, bias=False, device=device, dtype=dtype) self._layer_id = None self._dim = d_model self._hidden_dim = d_ff def forward(self, x: torch.Tensor) -> torch.Tensor: x = self.layer1(x) x = self.activation(x) x = self.layer2(x) return x def torch_attention_op(q_B_S_H_D: torch.Tensor, k_B_S_H_D: torch.Tensor, v_B_S_H_D: torch.Tensor, transformer_options: Optional[dict] = {}) -> torch.Tensor: """Computes multi-head attention using PyTorch's native implementation. This function provides a PyTorch backend alternative to Transformer Engine's attention operation. It rearranges the input tensors to match PyTorch's expected format, computes scaled dot-product attention, and rearranges the output back to the original format. The input tensor names use the following dimension conventions: - B: batch size - S: sequence length - H: number of attention heads - D: head dimension Args: q_B_S_H_D: Query tensor with shape (batch, seq_len, n_heads, head_dim) k_B_S_H_D: Key tensor with shape (batch, seq_len, n_heads, head_dim) v_B_S_H_D: Value tensor with shape (batch, seq_len, n_heads, head_dim) Returns: Attention output tensor with shape (batch, seq_len, n_heads * head_dim) """ in_q_shape = q_B_S_H_D.shape in_k_shape = k_B_S_H_D.shape q_B_H_S_D = rearrange(q_B_S_H_D, "b ... h k -> b h ... k").view(in_q_shape[0], in_q_shape[-2], -1, in_q_shape[-1]) k_B_H_S_D = rearrange(k_B_S_H_D, "b ... h v -> b h ... v").view(in_k_shape[0], in_k_shape[-2], -1, in_k_shape[-1]) v_B_H_S_D = rearrange(v_B_S_H_D, "b ... h v -> b h ... v").view(in_k_shape[0], in_k_shape[-2], -1, in_k_shape[-1]) return optimized_attention(q_B_H_S_D, k_B_H_S_D, v_B_H_S_D, in_q_shape[-2], skip_reshape=True, transformer_options=transformer_options) class Attention(nn.Module): """ A flexible attention module supporting both self-attention and cross-attention mechanisms. This module implements a multi-head attention layer that can operate in either self-attention or cross-attention mode. The mode is determined by whether a context dimension is provided. The implementation uses scaled dot-product attention and supports optional bias terms and dropout regularization. Args: query_dim (int): The dimensionality of the query vectors. context_dim (int, optional): The dimensionality of the context (key/value) vectors. If None, the module operates in self-attention mode using query_dim. Default: None n_heads (int, optional): Number of attention heads for multi-head attention. Default: 8 head_dim (int, optional): The dimension of each attention head. Default: 64 dropout (float, optional): Dropout probability applied to the output. Default: 0.0 qkv_format (str, optional): Format specification for QKV tensors. Default: "bshd" backend (str, optional): Backend to use for the attention operation. Default: "transformer_engine" Examples: >>> # Self-attention with 512 dimensions and 8 heads >>> self_attn = Attention(query_dim=512) >>> x = torch.randn(32, 16, 512) # (batch_size, seq_len, dim) >>> out = self_attn(x) # (32, 16, 512) >>> # Cross-attention >>> cross_attn = Attention(query_dim=512, context_dim=256) >>> query = torch.randn(32, 16, 512) >>> context = torch.randn(32, 8, 256) >>> out = cross_attn(query, context) # (32, 16, 512) """ def __init__( self, query_dim: int, context_dim: Optional[int] = None, n_heads: int = 8, head_dim: int = 64, dropout: float = 0.0, device=None, dtype=None, operations=None, ) -> None: super().__init__() logging.debug( f"Setting up {self.__class__.__name__}. Query dim is {query_dim}, context_dim is {context_dim} and using " f"{n_heads} heads with a dimension of {head_dim}." ) self.is_selfattn = context_dim is None # self attention context_dim = query_dim if context_dim is None else context_dim inner_dim = head_dim * n_heads self.n_heads = n_heads self.head_dim = head_dim self.query_dim = query_dim self.context_dim = context_dim self.q_proj = operations.Linear(query_dim, inner_dim, bias=False, device=device, dtype=dtype) self.q_norm = operations.RMSNorm(self.head_dim, eps=1e-6, device=device, dtype=dtype) self.k_proj = operations.Linear(context_dim, inner_dim, bias=False, device=device, dtype=dtype) self.k_norm = operations.RMSNorm(self.head_dim, eps=1e-6, device=device, dtype=dtype) self.v_proj = operations.Linear(context_dim, inner_dim, bias=False, device=device, dtype=dtype) self.v_norm = nn.Identity() self.output_proj = operations.Linear(inner_dim, query_dim, bias=False, device=device, dtype=dtype) self.output_dropout = nn.Dropout(dropout) if dropout > 1e-4 else nn.Identity() self.attn_op = torch_attention_op self._query_dim = query_dim self._context_dim = context_dim self._inner_dim = inner_dim def compute_qkv( self, x: torch.Tensor, context: Optional[torch.Tensor] = None, rope_emb: Optional[torch.Tensor] = None, ) -> tuple[torch.Tensor, torch.Tensor, torch.Tensor]: q = self.q_proj(x) context = x if context is None else context k = self.k_proj(context) v = self.v_proj(context) q, k, v = map( lambda t: rearrange(t, "b ... (h d) -> b ... h d", h=self.n_heads, d=self.head_dim), (q, k, v), ) def apply_norm_and_rotary_pos_emb( q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, rope_emb: Optional[torch.Tensor] ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: q = self.q_norm(q) k = self.k_norm(k) v = self.v_norm(v) if self.is_selfattn and rope_emb is not None: # only apply to self-attention! q = apply_rotary_pos_emb(q, rope_emb) k = apply_rotary_pos_emb(k, rope_emb) return q, k, v q, k, v = apply_norm_and_rotary_pos_emb(q, k, v, rope_emb) return q, k, v def compute_attention(self, q: torch.Tensor, k: torch.Tensor, v: torch.Tensor, transformer_options: Optional[dict] = {}) -> torch.Tensor: result = self.attn_op(q, k, v, transformer_options=transformer_options) # [B, S, H, D] return self.output_dropout(self.output_proj(result)) def forward( self, x: torch.Tensor, context: Optional[torch.Tensor] = None, rope_emb: Optional[torch.Tensor] = None, transformer_options: Optional[dict] = {}, ) -> torch.Tensor: """ Args: x (Tensor): The query tensor of shape [B, Mq, K] context (Optional[Tensor]): The key tensor of shape [B, Mk, K] or use x as context [self attention] if None """ q, k, v = self.compute_qkv(x, context, rope_emb=rope_emb) return self.compute_attention(q, k, v, transformer_options=transformer_options) class Timesteps(nn.Module): def __init__(self, num_channels: int): super().__init__() self.num_channels = num_channels def forward(self, timesteps_B_T: torch.Tensor) -> torch.Tensor: assert timesteps_B_T.ndim == 2, f"Expected 2D input, got {timesteps_B_T.ndim}" timesteps = timesteps_B_T.flatten().float() half_dim = self.num_channels // 2 exponent = -math.log(10000) * torch.arange(half_dim, dtype=torch.float32, device=timesteps.device) exponent = exponent / (half_dim - 0.0) emb = torch.exp(exponent) emb = timesteps[:, None].float() * emb[None, :] sin_emb = torch.sin(emb) cos_emb = torch.cos(emb) emb = torch.cat([cos_emb, sin_emb], dim=-1) return rearrange(emb, "(b t) d -> b t d", b=timesteps_B_T.shape[0], t=timesteps_B_T.shape[1]) class TimestepEmbedding(nn.Module): def __init__(self, in_features: int, out_features: int, use_adaln_lora: bool = False, device=None, dtype=None, operations=None): super().__init__() logging.debug( f"Using AdaLN LoRA Flag: {use_adaln_lora}. We enable bias if no AdaLN LoRA for backward compatibility." ) self.in_dim = in_features self.out_dim = out_features self.linear_1 = operations.Linear(in_features, out_features, bias=not use_adaln_lora, device=device, dtype=dtype) self.activation = nn.SiLU() self.use_adaln_lora = use_adaln_lora if use_adaln_lora: self.linear_2 = operations.Linear(out_features, 3 * out_features, bias=False, device=device, dtype=dtype) else: self.linear_2 = operations.Linear(out_features, out_features, bias=False, device=device, dtype=dtype) def forward(self, sample: torch.Tensor) -> Tuple[torch.Tensor, Optional[torch.Tensor]]: emb = self.linear_1(sample) emb = self.activation(emb) emb = self.linear_2(emb) if self.use_adaln_lora: adaln_lora_B_T_3D = emb emb_B_T_D = sample else: adaln_lora_B_T_3D = None emb_B_T_D = emb return emb_B_T_D, adaln_lora_B_T_3D class PatchEmbed(nn.Module): """ PatchEmbed is a module for embedding patches from an input tensor by applying either 3D or 2D convolutional layers, depending on the . This module can process inputs with temporal (video) and spatial (image) dimensions, making it suitable for video and image processing tasks. It supports dividing the input into patches and embedding each patch into a vector of size `out_channels`. Parameters: - spatial_patch_size (int): The size of each spatial patch. - temporal_patch_size (int): The size of each temporal patch. - in_channels (int): Number of input channels. Default: 3. - out_channels (int): The dimension of the embedding vector for each patch. Default: 768. - bias (bool): If True, adds a learnable bias to the output of the convolutional layers. Default: True. """ def __init__( self, spatial_patch_size: int, temporal_patch_size: int, in_channels: int = 3, out_channels: int = 768, device=None, dtype=None, operations=None ): super().__init__() self.spatial_patch_size = spatial_patch_size self.temporal_patch_size = temporal_patch_size self.proj = nn.Sequential( Rearrange( "b c (t r) (h m) (w n) -> b t h w (c r m n)", r=temporal_patch_size, m=spatial_patch_size, n=spatial_patch_size, ), operations.Linear( in_channels * spatial_patch_size * spatial_patch_size * temporal_patch_size, out_channels, bias=False, device=device, dtype=dtype ), ) self.dim = in_channels * spatial_patch_size * spatial_patch_size * temporal_patch_size def forward(self, x: torch.Tensor) -> torch.Tensor: """ Forward pass of the PatchEmbed module. Parameters: - x (torch.Tensor): The input tensor of shape (B, C, T, H, W) where B is the batch size, C is the number of channels, T is the temporal dimension, H is the height, and W is the width of the input. Returns: - torch.Tensor: The embedded patches as a tensor, with shape b t h w c. """ assert x.dim() == 5 _, _, T, H, W = x.shape assert ( H % self.spatial_patch_size == 0 and W % self.spatial_patch_size == 0 ), f"H,W {(H, W)} should be divisible by spatial_patch_size {self.spatial_patch_size}" assert T % self.temporal_patch_size == 0 x = self.proj(x) return x class FinalLayer(nn.Module): """ The final layer of video DiT. """ def __init__( self, hidden_size: int, spatial_patch_size: int, temporal_patch_size: int, out_channels: int, use_adaln_lora: bool = False, adaln_lora_dim: int = 256, device=None, dtype=None, operations=None ): super().__init__() self.layer_norm = operations.LayerNorm(hidden_size, elementwise_affine=False, eps=1e-6) self.linear = operations.Linear( hidden_size, spatial_patch_size * spatial_patch_size * temporal_patch_size * out_channels, bias=False, device=device, dtype=dtype ) self.hidden_size = hidden_size self.n_adaln_chunks = 2 self.use_adaln_lora = use_adaln_lora self.adaln_lora_dim = adaln_lora_dim if use_adaln_lora: self.adaln_modulation = nn.Sequential( nn.SiLU(), operations.Linear(hidden_size, adaln_lora_dim, bias=False, device=device, dtype=dtype), operations.Linear(adaln_lora_dim, self.n_adaln_chunks * hidden_size, bias=False, device=device, dtype=dtype), ) else: self.adaln_modulation = nn.Sequential( nn.SiLU(), operations.Linear(hidden_size, self.n_adaln_chunks * hidden_size, bias=False, device=device, dtype=dtype) ) def forward( self, x_B_T_H_W_D: torch.Tensor, emb_B_T_D: torch.Tensor, adaln_lora_B_T_3D: Optional[torch.Tensor] = None, ): if self.use_adaln_lora: assert adaln_lora_B_T_3D is not None shift_B_T_D, scale_B_T_D = ( self.adaln_modulation(emb_B_T_D) + adaln_lora_B_T_3D[:, :, : 2 * self.hidden_size] ).chunk(2, dim=-1) else: shift_B_T_D, scale_B_T_D = self.adaln_modulation(emb_B_T_D).chunk(2, dim=-1) shift_B_T_1_1_D, scale_B_T_1_1_D = rearrange(shift_B_T_D, "b t d -> b t 1 1 d"), rearrange( scale_B_T_D, "b t d -> b t 1 1 d" ) def _fn( _x_B_T_H_W_D: torch.Tensor, _norm_layer: nn.Module, _scale_B_T_1_1_D: torch.Tensor, _shift_B_T_1_1_D: torch.Tensor, ) -> torch.Tensor: return _norm_layer(_x_B_T_H_W_D) * (1 + _scale_B_T_1_1_D) + _shift_B_T_1_1_D x_B_T_H_W_D = _fn(x_B_T_H_W_D, self.layer_norm, scale_B_T_1_1_D, shift_B_T_1_1_D) x_B_T_H_W_O = self.linear(x_B_T_H_W_D) return x_B_T_H_W_O class Block(nn.Module): """ A transformer block that combines self-attention, cross-attention and MLP layers with AdaLN modulation. Each component (self-attention, cross-attention, MLP) has its own layer normalization and AdaLN modulation. Parameters: x_dim (int): Dimension of input features context_dim (int): Dimension of context features for cross-attention num_heads (int): Number of attention heads mlp_ratio (float): Multiplier for MLP hidden dimension. Default: 4.0 use_adaln_lora (bool): Whether to use AdaLN-LoRA modulation. Default: False adaln_lora_dim (int): Hidden dimension for AdaLN-LoRA layers. Default: 256 The block applies the following sequence: 1. Self-attention with AdaLN modulation 2. Cross-attention with AdaLN modulation 3. MLP with AdaLN modulation Each component uses skip connections and layer normalization. """ def __init__( self, x_dim: int, context_dim: int, num_heads: int, mlp_ratio: float = 4.0, use_adaln_lora: bool = False, adaln_lora_dim: int = 256, device=None, dtype=None, operations=None, ): super().__init__() self.x_dim = x_dim self.layer_norm_self_attn = operations.LayerNorm(x_dim, elementwise_affine=False, eps=1e-6, device=device, dtype=dtype) self.self_attn = Attention(x_dim, None, num_heads, x_dim // num_heads, device=device, dtype=dtype, operations=operations) self.layer_norm_cross_attn = operations.LayerNorm(x_dim, elementwise_affine=False, eps=1e-6, device=device, dtype=dtype) self.cross_attn = Attention( x_dim, context_dim, num_heads, x_dim // num_heads, device=device, dtype=dtype, operations=operations ) self.layer_norm_mlp = operations.LayerNorm(x_dim, elementwise_affine=False, eps=1e-6, device=device, dtype=dtype) self.mlp = GPT2FeedForward(x_dim, int(x_dim * mlp_ratio), device=device, dtype=dtype, operations=operations) self.use_adaln_lora = use_adaln_lora if self.use_adaln_lora: self.adaln_modulation_self_attn = nn.Sequential( nn.SiLU(), operations.Linear(x_dim, adaln_lora_dim, bias=False, device=device, dtype=dtype), operations.Linear(adaln_lora_dim, 3 * x_dim, bias=False, device=device, dtype=dtype), ) self.adaln_modulation_cross_attn = nn.Sequential( nn.SiLU(), operations.Linear(x_dim, adaln_lora_dim, bias=False, device=device, dtype=dtype), operations.Linear(adaln_lora_dim, 3 * x_dim, bias=False, device=device, dtype=dtype), ) self.adaln_modulation_mlp = nn.Sequential( nn.SiLU(), operations.Linear(x_dim, adaln_lora_dim, bias=False, device=device, dtype=dtype), operations.Linear(adaln_lora_dim, 3 * x_dim, bias=False, device=device, dtype=dtype), ) else: self.adaln_modulation_self_attn = nn.Sequential(nn.SiLU(), operations.Linear(x_dim, 3 * x_dim, bias=False, device=device, dtype=dtype)) self.adaln_modulation_cross_attn = nn.Sequential(nn.SiLU(), operations.Linear(x_dim, 3 * x_dim, bias=False, device=device, dtype=dtype)) self.adaln_modulation_mlp = nn.Sequential(nn.SiLU(), operations.Linear(x_dim, 3 * x_dim, bias=False, device=device, dtype=dtype)) def forward( self, x_B_T_H_W_D: torch.Tensor, emb_B_T_D: torch.Tensor, crossattn_emb: torch.Tensor, rope_emb_L_1_1_D: Optional[torch.Tensor] = None, adaln_lora_B_T_3D: Optional[torch.Tensor] = None, extra_per_block_pos_emb: Optional[torch.Tensor] = None, transformer_options: Optional[dict] = {}, ) -> torch.Tensor: residual_dtype = x_B_T_H_W_D.dtype compute_dtype = emb_B_T_D.dtype if extra_per_block_pos_emb is not None: x_B_T_H_W_D = x_B_T_H_W_D + extra_per_block_pos_emb if self.use_adaln_lora: shift_self_attn_B_T_D, scale_self_attn_B_T_D, gate_self_attn_B_T_D = ( self.adaln_modulation_self_attn(emb_B_T_D) + adaln_lora_B_T_3D ).chunk(3, dim=-1) shift_cross_attn_B_T_D, scale_cross_attn_B_T_D, gate_cross_attn_B_T_D = ( self.adaln_modulation_cross_attn(emb_B_T_D) + adaln_lora_B_T_3D ).chunk(3, dim=-1) shift_mlp_B_T_D, scale_mlp_B_T_D, gate_mlp_B_T_D = ( self.adaln_modulation_mlp(emb_B_T_D) + adaln_lora_B_T_3D ).chunk(3, dim=-1) else: shift_self_attn_B_T_D, scale_self_attn_B_T_D, gate_self_attn_B_T_D = self.adaln_modulation_self_attn( emb_B_T_D ).chunk(3, dim=-1) shift_cross_attn_B_T_D, scale_cross_attn_B_T_D, gate_cross_attn_B_T_D = self.adaln_modulation_cross_attn( emb_B_T_D ).chunk(3, dim=-1) shift_mlp_B_T_D, scale_mlp_B_T_D, gate_mlp_B_T_D = self.adaln_modulation_mlp(emb_B_T_D).chunk(3, dim=-1) # Reshape tensors from (B, T, D) to (B, T, 1, 1, D) for broadcasting shift_self_attn_B_T_1_1_D = rearrange(shift_self_attn_B_T_D, "b t d -> b t 1 1 d") scale_self_attn_B_T_1_1_D = rearrange(scale_self_attn_B_T_D, "b t d -> b t 1 1 d") gate_self_attn_B_T_1_1_D = rearrange(gate_self_attn_B_T_D, "b t d -> b t 1 1 d") shift_cross_attn_B_T_1_1_D = rearrange(shift_cross_attn_B_T_D, "b t d -> b t 1 1 d") scale_cross_attn_B_T_1_1_D = rearrange(scale_cross_attn_B_T_D, "b t d -> b t 1 1 d") gate_cross_attn_B_T_1_1_D = rearrange(gate_cross_attn_B_T_D, "b t d -> b t 1 1 d") shift_mlp_B_T_1_1_D = rearrange(shift_mlp_B_T_D, "b t d -> b t 1 1 d") scale_mlp_B_T_1_1_D = rearrange(scale_mlp_B_T_D, "b t d -> b t 1 1 d") gate_mlp_B_T_1_1_D = rearrange(gate_mlp_B_T_D, "b t d -> b t 1 1 d") B, T, H, W, D = x_B_T_H_W_D.shape def _fn(_x_B_T_H_W_D, _norm_layer, _scale_B_T_1_1_D, _shift_B_T_1_1_D): return _norm_layer(_x_B_T_H_W_D) * (1 + _scale_B_T_1_1_D) + _shift_B_T_1_1_D normalized_x_B_T_H_W_D = _fn( x_B_T_H_W_D, self.layer_norm_self_attn, scale_self_attn_B_T_1_1_D, shift_self_attn_B_T_1_1_D, ) result_B_T_H_W_D = rearrange( self.self_attn( # normalized_x_B_T_HW_D, rearrange(normalized_x_B_T_H_W_D.to(compute_dtype), "b t h w d -> b (t h w) d"), None, rope_emb=rope_emb_L_1_1_D, transformer_options=transformer_options, ), "b (t h w) d -> b t h w d", t=T, h=H, w=W, ) x_B_T_H_W_D = x_B_T_H_W_D + gate_self_attn_B_T_1_1_D.to(residual_dtype) * result_B_T_H_W_D.to(residual_dtype) def _x_fn( _x_B_T_H_W_D: torch.Tensor, layer_norm_cross_attn: Callable, _scale_cross_attn_B_T_1_1_D: torch.Tensor, _shift_cross_attn_B_T_1_1_D: torch.Tensor, transformer_options: Optional[dict] = {}, ) -> torch.Tensor: _normalized_x_B_T_H_W_D = _fn( _x_B_T_H_W_D, layer_norm_cross_attn, _scale_cross_attn_B_T_1_1_D, _shift_cross_attn_B_T_1_1_D ) _result_B_T_H_W_D = rearrange( self.cross_attn( rearrange(_normalized_x_B_T_H_W_D.to(compute_dtype), "b t h w d -> b (t h w) d"), crossattn_emb, rope_emb=rope_emb_L_1_1_D, transformer_options=transformer_options, ), "b (t h w) d -> b t h w d", t=T, h=H, w=W, ) return _result_B_T_H_W_D result_B_T_H_W_D = _x_fn( x_B_T_H_W_D, self.layer_norm_cross_attn, scale_cross_attn_B_T_1_1_D, shift_cross_attn_B_T_1_1_D, transformer_options=transformer_options, ) x_B_T_H_W_D = result_B_T_H_W_D.to(residual_dtype) * gate_cross_attn_B_T_1_1_D.to(residual_dtype) + x_B_T_H_W_D normalized_x_B_T_H_W_D = _fn( x_B_T_H_W_D, self.layer_norm_mlp, scale_mlp_B_T_1_1_D, shift_mlp_B_T_1_1_D, ) result_B_T_H_W_D = self.mlp(normalized_x_B_T_H_W_D.to(compute_dtype)) x_B_T_H_W_D = x_B_T_H_W_D + gate_mlp_B_T_1_1_D.to(residual_dtype) * result_B_T_H_W_D.to(residual_dtype) return x_B_T_H_W_D class MiniTrainDIT(nn.Module): """ A clean impl of DIT that can load and reproduce the training results of the original DIT model in~(cosmos 1) A general implementation of adaln-modulated VIT-like~(DiT) transformer for video processing. Args: max_img_h (int): Maximum height of the input images. max_img_w (int): Maximum width of the input images. max_frames (int): Maximum number of frames in the video sequence. in_channels (int): Number of input channels (e.g., RGB channels for color images). out_channels (int): Number of output channels. patch_spatial (tuple): Spatial resolution of patches for input processing. patch_temporal (int): Temporal resolution of patches for input processing. concat_padding_mask (bool): If True, includes a mask channel in the input to handle padding. model_channels (int): Base number of channels used throughout the model. num_blocks (int): Number of transformer blocks. num_heads (int): Number of heads in the multi-head attention layers. mlp_ratio (float): Expansion ratio for MLP blocks. crossattn_emb_channels (int): Number of embedding channels for cross-attention. pos_emb_cls (str): Type of positional embeddings. pos_emb_learnable (bool): Whether positional embeddings are learnable. pos_emb_interpolation (str): Method for interpolating positional embeddings. min_fps (int): Minimum frames per second. max_fps (int): Maximum frames per second. use_adaln_lora (bool): Whether to use AdaLN-LoRA. adaln_lora_dim (int): Dimension for AdaLN-LoRA. rope_h_extrapolation_ratio (float): Height extrapolation ratio for RoPE. rope_w_extrapolation_ratio (float): Width extrapolation ratio for RoPE. rope_t_extrapolation_ratio (float): Temporal extrapolation ratio for RoPE. extra_per_block_abs_pos_emb (bool): Whether to use extra per-block absolute positional embeddings. extra_h_extrapolation_ratio (float): Height extrapolation ratio for extra embeddings. extra_w_extrapolation_ratio (float): Width extrapolation ratio for extra embeddings. extra_t_extrapolation_ratio (float): Temporal extrapolation ratio for extra embeddings. """ def __init__( self, max_img_h: int, max_img_w: int, max_frames: int, in_channels: int, out_channels: int, patch_spatial: int, # tuple, patch_temporal: int, concat_padding_mask: bool = True, # attention settings model_channels: int = 768, num_blocks: int = 10, num_heads: int = 16, mlp_ratio: float = 4.0, # cross attention settings crossattn_emb_channels: int = 1024, # positional embedding settings pos_emb_cls: str = "sincos", pos_emb_learnable: bool = False, pos_emb_interpolation: str = "crop", min_fps: int = 1, max_fps: int = 30, use_adaln_lora: bool = False, adaln_lora_dim: int = 256, rope_h_extrapolation_ratio: float = 1.0, rope_w_extrapolation_ratio: float = 1.0, rope_t_extrapolation_ratio: float = 1.0, extra_per_block_abs_pos_emb: bool = False, extra_h_extrapolation_ratio: float = 1.0, extra_w_extrapolation_ratio: float = 1.0, extra_t_extrapolation_ratio: float = 1.0, rope_enable_fps_modulation: bool = True, image_model=None, device=None, dtype=None, operations=None, ) -> None: super().__init__() self.dtype = dtype self.max_img_h = max_img_h self.max_img_w = max_img_w self.max_frames = max_frames self.in_channels = in_channels self.out_channels = out_channels self.patch_spatial = patch_spatial self.patch_temporal = patch_temporal self.num_heads = num_heads self.num_blocks = num_blocks self.model_channels = model_channels self.concat_padding_mask = concat_padding_mask # positional embedding settings self.pos_emb_cls = pos_emb_cls self.pos_emb_learnable = pos_emb_learnable self.pos_emb_interpolation = pos_emb_interpolation self.min_fps = min_fps self.max_fps = max_fps self.rope_h_extrapolation_ratio = rope_h_extrapolation_ratio self.rope_w_extrapolation_ratio = rope_w_extrapolation_ratio self.rope_t_extrapolation_ratio = rope_t_extrapolation_ratio self.extra_per_block_abs_pos_emb = extra_per_block_abs_pos_emb self.extra_h_extrapolation_ratio = extra_h_extrapolation_ratio self.extra_w_extrapolation_ratio = extra_w_extrapolation_ratio self.extra_t_extrapolation_ratio = extra_t_extrapolation_ratio self.rope_enable_fps_modulation = rope_enable_fps_modulation self.build_pos_embed(device=device, dtype=dtype) self.use_adaln_lora = use_adaln_lora self.adaln_lora_dim = adaln_lora_dim self.t_embedder = nn.Sequential( Timesteps(model_channels), TimestepEmbedding(model_channels, model_channels, use_adaln_lora=use_adaln_lora, device=device, dtype=dtype, operations=operations,), ) in_channels = in_channels + 1 if concat_padding_mask else in_channels self.x_embedder = PatchEmbed( spatial_patch_size=patch_spatial, temporal_patch_size=patch_temporal, in_channels=in_channels, out_channels=model_channels, device=device, dtype=dtype, operations=operations, ) self.blocks = nn.ModuleList( [ Block( x_dim=model_channels, context_dim=crossattn_emb_channels, num_heads=num_heads, mlp_ratio=mlp_ratio, use_adaln_lora=use_adaln_lora, adaln_lora_dim=adaln_lora_dim, device=device, dtype=dtype, operations=operations, ) for _ in range(num_blocks) ] ) self.final_layer = FinalLayer( hidden_size=self.model_channels, spatial_patch_size=self.patch_spatial, temporal_patch_size=self.patch_temporal, out_channels=self.out_channels, use_adaln_lora=self.use_adaln_lora, adaln_lora_dim=self.adaln_lora_dim, device=device, dtype=dtype, operations=operations, ) self.t_embedding_norm = operations.RMSNorm(model_channels, eps=1e-6, device=device, dtype=dtype) def build_pos_embed(self, device=None, dtype=None) -> None: if self.pos_emb_cls == "rope3d": cls_type = VideoRopePosition3DEmb else: raise ValueError(f"Unknown pos_emb_cls {self.pos_emb_cls}") logging.debug(f"Building positional embedding with {self.pos_emb_cls} class, impl {cls_type}") kwargs = dict( model_channels=self.model_channels, len_h=self.max_img_h // self.patch_spatial, len_w=self.max_img_w // self.patch_spatial, len_t=self.max_frames // self.patch_temporal, max_fps=self.max_fps, min_fps=self.min_fps, is_learnable=self.pos_emb_learnable, interpolation=self.pos_emb_interpolation, head_dim=self.model_channels // self.num_heads, h_extrapolation_ratio=self.rope_h_extrapolation_ratio, w_extrapolation_ratio=self.rope_w_extrapolation_ratio, t_extrapolation_ratio=self.rope_t_extrapolation_ratio, enable_fps_modulation=self.rope_enable_fps_modulation, device=device, ) self.pos_embedder = cls_type( **kwargs, # type: ignore ) if self.extra_per_block_abs_pos_emb: kwargs["h_extrapolation_ratio"] = self.extra_h_extrapolation_ratio kwargs["w_extrapolation_ratio"] = self.extra_w_extrapolation_ratio kwargs["t_extrapolation_ratio"] = self.extra_t_extrapolation_ratio kwargs["device"] = device kwargs["dtype"] = dtype self.extra_pos_embedder = LearnablePosEmbAxis( **kwargs, # type: ignore ) def prepare_embedded_sequence( self, x_B_C_T_H_W: torch.Tensor, fps: Optional[torch.Tensor] = None, padding_mask: Optional[torch.Tensor] = None, ) -> Tuple[torch.Tensor, Optional[torch.Tensor], Optional[torch.Tensor]]: """ Prepares an embedded sequence tensor by applying positional embeddings and handling padding masks. Args: x_B_C_T_H_W (torch.Tensor): video fps (Optional[torch.Tensor]): Frames per second tensor to be used for positional embedding when required. If None, a default value (`self.base_fps`) will be used. padding_mask (Optional[torch.Tensor]): current it is not used Returns: Tuple[torch.Tensor, Optional[torch.Tensor]]: - A tensor of shape (B, T, H, W, D) with the embedded sequence. - An optional positional embedding tensor, returned only if the positional embedding class (`self.pos_emb_cls`) includes 'rope'. Otherwise, None. Notes: - If `self.concat_padding_mask` is True, a padding mask channel is concatenated to the input tensor. - The method of applying positional embeddings depends on the value of `self.pos_emb_cls`. - If 'rope' is in `self.pos_emb_cls` (case insensitive), the positional embeddings are generated using the `self.pos_embedder` with the shape [T, H, W]. - If "fps_aware" is in `self.pos_emb_cls`, the positional embeddings are generated using the `self.pos_embedder` with the fps tensor. - Otherwise, the positional embeddings are generated without considering fps. """ if self.concat_padding_mask: if padding_mask is None: padding_mask = torch.zeros(x_B_C_T_H_W.shape[0], 1, x_B_C_T_H_W.shape[3], x_B_C_T_H_W.shape[4], dtype=x_B_C_T_H_W.dtype, device=x_B_C_T_H_W.device) else: padding_mask = transforms.functional.resize( padding_mask, list(x_B_C_T_H_W.shape[-2:]), interpolation=transforms.InterpolationMode.NEAREST ) x_B_C_T_H_W = torch.cat( [x_B_C_T_H_W, padding_mask.unsqueeze(1).repeat(1, 1, x_B_C_T_H_W.shape[2], 1, 1)], dim=1 ) x_B_T_H_W_D = self.x_embedder(x_B_C_T_H_W) if self.extra_per_block_abs_pos_emb: extra_pos_emb = self.extra_pos_embedder(x_B_T_H_W_D, fps=fps, device=x_B_C_T_H_W.device, dtype=x_B_C_T_H_W.dtype) else: extra_pos_emb = None if "rope" in self.pos_emb_cls.lower(): return x_B_T_H_W_D, self.pos_embedder(x_B_T_H_W_D, fps=fps, device=x_B_C_T_H_W.device), extra_pos_emb x_B_T_H_W_D = x_B_T_H_W_D + self.pos_embedder(x_B_T_H_W_D, device=x_B_C_T_H_W.device) # [B, T, H, W, D] return x_B_T_H_W_D, None, extra_pos_emb def unpatchify(self, x_B_T_H_W_M: torch.Tensor) -> torch.Tensor: x_B_C_Tt_Hp_Wp = rearrange( x_B_T_H_W_M, "B T H W (p1 p2 t C) -> B C (T t) (H p1) (W p2)", p1=self.patch_spatial, p2=self.patch_spatial, t=self.patch_temporal, ) return x_B_C_Tt_Hp_Wp def forward(self, x: torch.Tensor, timesteps: torch.Tensor, context: torch.Tensor, fps: Optional[torch.Tensor] = None, padding_mask: Optional[torch.Tensor] = None, **kwargs, ): return comfy.patcher_extension.WrapperExecutor.new_class_executor( self._forward, self, comfy.patcher_extension.get_all_wrappers(comfy.patcher_extension.WrappersMP.DIFFUSION_MODEL, kwargs.get("transformer_options", {})) ).execute(x, timesteps, context, fps, padding_mask, **kwargs) def _forward( self, x: torch.Tensor, timesteps: torch.Tensor, context: torch.Tensor, fps: Optional[torch.Tensor] = None, padding_mask: Optional[torch.Tensor] = None, **kwargs, ): orig_shape = list(x.shape) x = comfy.ldm.common_dit.pad_to_patch_size(x, (self.patch_temporal, self.patch_spatial, self.patch_spatial)) x_B_C_T_H_W = x timesteps_B_T = timesteps crossattn_emb = context """ Args: x: (B, C, T, H, W) tensor of spatial-temp inputs timesteps: (B, ) tensor of timesteps crossattn_emb: (B, N, D) tensor of cross-attention embeddings """ x_B_T_H_W_D, rope_emb_L_1_1_D, extra_pos_emb_B_T_H_W_D_or_T_H_W_B_D = self.prepare_embedded_sequence( x_B_C_T_H_W, fps=fps, padding_mask=padding_mask, ) if timesteps_B_T.ndim == 1: timesteps_B_T = timesteps_B_T.unsqueeze(1) t_embedding_B_T_D, adaln_lora_B_T_3D = self.t_embedder[1](self.t_embedder[0](timesteps_B_T).to(x_B_T_H_W_D.dtype)) t_embedding_B_T_D = self.t_embedding_norm(t_embedding_B_T_D) # for logging purpose affline_scale_log_info = {} affline_scale_log_info["t_embedding_B_T_D"] = t_embedding_B_T_D.detach() self.affline_scale_log_info = affline_scale_log_info self.affline_emb = t_embedding_B_T_D self.crossattn_emb = crossattn_emb if extra_pos_emb_B_T_H_W_D_or_T_H_W_B_D is not None: assert ( x_B_T_H_W_D.shape == extra_pos_emb_B_T_H_W_D_or_T_H_W_B_D.shape ), f"{x_B_T_H_W_D.shape} != {extra_pos_emb_B_T_H_W_D_or_T_H_W_B_D.shape}" block_kwargs = { "rope_emb_L_1_1_D": rope_emb_L_1_1_D.unsqueeze(1).unsqueeze(0), "adaln_lora_B_T_3D": adaln_lora_B_T_3D, "extra_per_block_pos_emb": extra_pos_emb_B_T_H_W_D_or_T_H_W_B_D, "transformer_options": kwargs.get("transformer_options", {}), } # The residual stream for this model has large values. To make fp16 compute_dtype work, we keep the residual stream # in fp32, but run attention and MLP modules in fp16. # An alternate method that clamps fp16 values "works" in the sense that it makes coherent images, but there is noticeable # quality degradation and visual artifacts. if x_B_T_H_W_D.dtype == torch.float16: x_B_T_H_W_D = x_B_T_H_W_D.float() for block in self.blocks: x_B_T_H_W_D = block( x_B_T_H_W_D, t_embedding_B_T_D, crossattn_emb, **block_kwargs, ) x_B_T_H_W_O = self.final_layer(x_B_T_H_W_D.to(crossattn_emb.dtype), t_embedding_B_T_D, adaln_lora_B_T_3D=adaln_lora_B_T_3D) x_B_C_Tt_Hp_Wp = self.unpatchify(x_B_T_H_W_O)[:, :, :orig_shape[-3], :orig_shape[-2], :orig_shape[-1]] return x_B_C_Tt_Hp_Wp
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/ldm/cosmos/predict2.py", "license": "GNU General Public License v3.0", "lines": 787, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:alembic_db/env.py
from sqlalchemy import engine_from_config from sqlalchemy import pool from alembic import context # this is the Alembic Config object, which provides # access to the values within the .ini file in use. config = context.config from app.database.models import Base target_metadata = Base.metadata # other values from the config, defined by the needs of env.py, # can be acquired: # my_important_option = config.get_main_option("my_important_option") # ... etc. def run_migrations_offline() -> None: """Run migrations in 'offline' mode. This configures the context with just a URL and not an Engine, though an Engine is acceptable here as well. By skipping the Engine creation we don't even need a DBAPI to be available. Calls to context.execute() here emit the given string to the script output. """ url = config.get_main_option("sqlalchemy.url") context.configure( url=url, target_metadata=target_metadata, literal_binds=True, dialect_opts={"paramstyle": "named"}, ) with context.begin_transaction(): context.run_migrations() def run_migrations_online() -> None: """Run migrations in 'online' mode. In this scenario we need to create an Engine and associate a connection with the context. """ connectable = engine_from_config( config.get_section(config.config_ini_section, {}), prefix="sqlalchemy.", poolclass=pool.NullPool, ) with connectable.connect() as connection: context.configure( connection=connection, target_metadata=target_metadata ) with context.begin_transaction(): context.run_migrations() if context.is_offline_mode(): run_migrations_offline() else: run_migrations_online()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "alembic_db/env.py", "license": "GNU General Public License v3.0", "lines": 50, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:app/database/db.py
import logging import os import shutil from app.logger import log_startup_warning from utils.install_util import get_missing_requirements_message from comfy.cli_args import args _DB_AVAILABLE = False Session = None try: from alembic import command from alembic.config import Config from alembic.runtime.migration import MigrationContext from alembic.script import ScriptDirectory from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker _DB_AVAILABLE = True except ImportError as e: log_startup_warning( f""" ------------------------------------------------------------------------ Error importing dependencies: {e} {get_missing_requirements_message()} This error is happening because ComfyUI now uses a local sqlite database. ------------------------------------------------------------------------ """.strip() ) def dependencies_available(): """ Temporary function to check if the dependencies are available """ return _DB_AVAILABLE def can_create_session(): """ Temporary function to check if the database is available to create a session During initial release there may be environmental issues (or missing dependencies) that prevent the database from being created """ return dependencies_available() and Session is not None def get_alembic_config(): root_path = os.path.join(os.path.dirname(__file__), "../..") config_path = os.path.abspath(os.path.join(root_path, "alembic.ini")) scripts_path = os.path.abspath(os.path.join(root_path, "alembic_db")) config = Config(config_path) config.set_main_option("script_location", scripts_path) config.set_main_option("sqlalchemy.url", args.database_url) return config def get_db_path(): url = args.database_url if url.startswith("sqlite:///"): return url.split("///")[1] else: raise ValueError(f"Unsupported database URL '{url}'.") def init_db(): db_url = args.database_url logging.debug(f"Database URL: {db_url}") db_path = get_db_path() db_exists = os.path.exists(db_path) config = get_alembic_config() # Check if we need to upgrade engine = create_engine(db_url) conn = engine.connect() context = MigrationContext.configure(conn) current_rev = context.get_current_revision() script = ScriptDirectory.from_config(config) target_rev = script.get_current_head() if target_rev is None: logging.warning("No target revision found.") elif current_rev != target_rev: # Backup the database pre upgrade backup_path = db_path + ".bkp" if db_exists: shutil.copy(db_path, backup_path) else: backup_path = None try: command.upgrade(config, target_rev) logging.info(f"Database upgraded from {current_rev} to {target_rev}") except Exception as e: if backup_path: # Restore the database from backup if upgrade fails shutil.copy(backup_path, db_path) os.remove(backup_path) logging.exception("Error upgrading database: ") raise e global Session Session = sessionmaker(bind=engine) def create_session(): return Session()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "app/database/db.py", "license": "GNU General Public License v3.0", "lines": 87, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:app/database/models.py
from typing import Any from datetime import datetime from sqlalchemy.orm import DeclarativeBase class Base(DeclarativeBase): pass def to_dict(obj: Any, include_none: bool = False) -> dict[str, Any]: fields = obj.__table__.columns.keys() out: dict[str, Any] = {} for field in fields: val = getattr(obj, field) if val is None and not include_none: continue if isinstance(val, datetime): out[field] = val.isoformat() else: out[field] = val return out # TODO: Define models here
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "app/database/models.py", "license": "GNU General Public License v3.0", "lines": 18, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:utils/install_util.py
from pathlib import Path import sys import logging import re # The path to the requirements.txt file requirements_path = Path(__file__).parents[1] / "requirements.txt" def get_missing_requirements_message(): """The warning message to display when a package is missing.""" extra = "" if sys.flags.no_user_site: extra = "-s " return f""" Please install the updated requirements.txt file by running: {sys.executable} {extra}-m pip install -r {requirements_path} If you are on the portable package you can run: update\\update_comfyui.bat to solve this problem. """.strip() def is_valid_version(version: str) -> bool: """Validate if a string is a valid semantic version (X.Y.Z format).""" pattern = r"^(\d+)\.(\d+)\.(\d+)$" return bool(re.match(pattern, version)) PACKAGE_VERSIONS = {} def get_required_packages_versions(): if len(PACKAGE_VERSIONS) > 0: return PACKAGE_VERSIONS.copy() out = PACKAGE_VERSIONS try: with open(requirements_path, "r", encoding="utf-8") as f: for line in f: line = line.strip().replace(">=", "==") s = line.split("==") if len(s) == 2: version_str = s[-1] if not is_valid_version(version_str): logging.error(f"Invalid version format in requirements.txt: {version_str}") continue out[s[0]] = version_str return out.copy() except FileNotFoundError: logging.error("requirements.txt not found.") return None except Exception as e: logging.error(f"Error reading requirements.txt: {e}") return None
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "utils/install_util.py", "license": "GNU General Public License v3.0", "lines": 43, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy_config/config_parser.py
import os from pathlib import Path from typing import Optional from pydantic_settings import PydanticBaseSettingsSource, TomlConfigSettingsSource from comfy_config.types import ( ComfyConfig, ProjectConfig, PyProjectConfig, PyProjectSettings ) def validate_and_extract_os_classifiers(classifiers: list) -> list: os_classifiers = [c for c in classifiers if c.startswith("Operating System :: ")] if not os_classifiers: return [] os_values = [c[len("Operating System :: ") :] for c in os_classifiers] valid_os_prefixes = {"Microsoft", "POSIX", "MacOS", "OS Independent"} for os_value in os_values: if not any(os_value.startswith(prefix) for prefix in valid_os_prefixes): return [] return os_values def validate_and_extract_accelerator_classifiers(classifiers: list) -> list: accelerator_classifiers = [c for c in classifiers if c.startswith("Environment ::")] if not accelerator_classifiers: return [] accelerator_values = [c[len("Environment :: ") :] for c in accelerator_classifiers] valid_accelerators = { "GPU :: NVIDIA CUDA", "GPU :: AMD ROCm", "GPU :: Intel Arc", "NPU :: Huawei Ascend", "GPU :: Apple Metal", } for accelerator_value in accelerator_values: if accelerator_value not in valid_accelerators: return [] return accelerator_values """ Extract configuration from a custom node directory's pyproject.toml file or a Python file. This function reads and parses the pyproject.toml file in the specified directory to extract project and ComfyUI-specific configuration information. If no pyproject.toml file is found, it creates a minimal configuration using the folder name as the project name. If a Python file is provided, it uses the file name (without extension) as the project name. Args: path (str): Path to the directory containing the pyproject.toml file, or path to a .py file. If pyproject.toml doesn't exist in a directory, the folder name will be used as the default project name. If a .py file is provided, the filename (without .py extension) will be used as the project name. Returns: Optional[PyProjectConfig]: A PyProjectConfig object containing: - project: Basic project information (name, version, dependencies, etc.) - tool_comfy: ComfyUI-specific configuration (publisher_id, models, etc.) Returns None if configuration extraction fails or if the provided file is not a Python file. Notes: - If pyproject.toml is missing in a directory, creates a default config with folder name - If a .py file is provided, creates a default config with filename (without extension) - Returns None for non-Python files Example: >>> from comfy_config import config_parser >>> # For directory >>> custom_node_dir = os.path.dirname(os.path.realpath(__file__)) >>> project_config = config_parser.extract_node_configuration(custom_node_dir) >>> print(project_config.project.name) # "my_custom_node" or name from pyproject.toml >>> >>> # For single-file Python node file >>> py_file_path = os.path.realpath(__file__) # "/path/to/my_node.py" >>> project_config = config_parser.extract_node_configuration(py_file_path) >>> print(project_config.project.name) # "my_node" """ def extract_node_configuration(path) -> Optional[PyProjectConfig]: if os.path.isfile(path): file_path = Path(path) if file_path.suffix.lower() != '.py': return None project_name = file_path.stem project = ProjectConfig(name=project_name) comfy = ComfyConfig() return PyProjectConfig(project=project, tool_comfy=comfy) folder_name = os.path.basename(path) toml_path = Path(path) / "pyproject.toml" if not toml_path.exists(): project = ProjectConfig(name=folder_name) comfy = ComfyConfig() return PyProjectConfig(project=project, tool_comfy=comfy) raw_settings = load_pyproject_settings(toml_path) project_data = raw_settings.project tool_data = raw_settings.tool comfy_data = tool_data.get("comfy", {}) if tool_data else {} dependencies = project_data.get("dependencies", []) supported_comfyui_frontend_version = "" for dep in dependencies: if isinstance(dep, str) and dep.startswith("comfyui-frontend-package"): supported_comfyui_frontend_version = dep.removeprefix("comfyui-frontend-package") break supported_comfyui_version = comfy_data.get("requires-comfyui", "") classifiers = project_data.get('classifiers', []) supported_os = validate_and_extract_os_classifiers(classifiers) supported_accelerators = validate_and_extract_accelerator_classifiers(classifiers) project_data['supported_os'] = supported_os project_data['supported_accelerators'] = supported_accelerators project_data['supported_comfyui_frontend_version'] = supported_comfyui_frontend_version project_data['supported_comfyui_version'] = supported_comfyui_version return PyProjectConfig(project=project_data, tool_comfy=comfy_data) def load_pyproject_settings(toml_path: Path) -> PyProjectSettings: class PyProjectLoader(PyProjectSettings): @classmethod def settings_customise_sources( cls, settings_cls, init_settings: PydanticBaseSettingsSource, env_settings: PydanticBaseSettingsSource, dotenv_settings: PydanticBaseSettingsSource, file_secret_settings: PydanticBaseSettingsSource, ): return (TomlConfigSettingsSource(settings_cls, toml_path),) return PyProjectLoader()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_config/config_parser.py", "license": "GNU General Public License v3.0", "lines": 118, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_config/types.py
from pydantic import BaseModel, Field, field_validator from pydantic_settings import BaseSettings, SettingsConfigDict from typing import List, Optional # IMPORTANT: The type definitions specified in pyproject.toml for custom nodes # must remain synchronized with the corresponding files in the https://github.com/Comfy-Org/comfy-cli/blob/main/comfy_cli/registry/types.py. # Any changes to one must be reflected in the other to maintain consistency. class NodeVersion(BaseModel): changelog: str dependencies: List[str] deprecated: bool id: str version: str download_url: str class Node(BaseModel): id: str name: str description: str author: Optional[str] = None license: Optional[str] = None icon: Optional[str] = None repository: Optional[str] = None tags: List[str] = Field(default_factory=list) latest_version: Optional[NodeVersion] = None class PublishNodeVersionResponse(BaseModel): node_version: NodeVersion signedUrl: str class URLs(BaseModel): homepage: str = Field(default="", alias="Homepage") documentation: str = Field(default="", alias="Documentation") repository: str = Field(default="", alias="Repository") issues: str = Field(default="", alias="Issues") class Model(BaseModel): location: str model_url: str class ComfyConfig(BaseModel): publisher_id: str = Field(default="", alias="PublisherId") display_name: str = Field(default="", alias="DisplayName") icon: str = Field(default="", alias="Icon") models: List[Model] = Field(default_factory=list, alias="Models") includes: List[str] = Field(default_factory=list) web: Optional[str] = None banner_url: str = "" class License(BaseModel): file: str = "" text: str = "" class ProjectConfig(BaseModel): name: str = "" description: str = "" version: str = "1.0.0" requires_python: str = Field(default=">= 3.9", alias="requires-python") dependencies: List[str] = Field(default_factory=list) license: License = Field(default_factory=License) urls: URLs = Field(default_factory=URLs) supported_os: List[str] = Field(default_factory=list) supported_accelerators: List[str] = Field(default_factory=list) supported_comfyui_version: str = "" supported_comfyui_frontend_version: str = "" @field_validator('license', mode='before') @classmethod def validate_license(cls, v): if isinstance(v, str): return License(text=v) elif isinstance(v, dict): return License(**v) elif isinstance(v, License): return v else: return License() class PyProjectConfig(BaseModel): project: ProjectConfig = Field(default_factory=ProjectConfig) tool_comfy: ComfyConfig = Field(default_factory=ComfyConfig) class PyProjectSettings(BaseSettings): project: dict = Field(default_factory=dict) tool: dict = Field(default_factory=dict) model_config = SettingsConfigDict(extra='allow')
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_config/types.py", "license": "GNU General Public License v3.0", "lines": 75, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:tests-unit/comfy_extras_test/image_stitch_test.py
import torch from unittest.mock import patch, MagicMock # Mock nodes module to prevent CUDA initialization during import mock_nodes = MagicMock() mock_nodes.MAX_RESOLUTION = 16384 # Mock server module for PromptServer mock_server = MagicMock() with patch.dict('sys.modules', {'nodes': mock_nodes, 'server': mock_server}): from comfy_extras.nodes_images import ImageStitch class TestImageStitch: def create_test_image(self, batch_size=1, height=64, width=64, channels=3): """Helper to create test images with specific dimensions""" return torch.rand(batch_size, height, width, channels) def test_no_image2_passthrough(self): """Test that when image2 is None, image1 is returned unchanged""" node = ImageStitch() image1 = self.create_test_image() result = node.stitch(image1, "right", True, 0, "white", image2=None) assert len(result.result) == 1 assert torch.equal(result[0], image1) def test_basic_horizontal_stitch_right(self): """Test basic horizontal stitching to the right""" node = ImageStitch() image1 = self.create_test_image(height=32, width=32) image2 = self.create_test_image(height=32, width=24) result = node.stitch(image1, "right", False, 0, "white", image2) assert result[0].shape == (1, 32, 56, 3) # 32 + 24 width def test_basic_horizontal_stitch_left(self): """Test basic horizontal stitching to the left""" node = ImageStitch() image1 = self.create_test_image(height=32, width=32) image2 = self.create_test_image(height=32, width=24) result = node.stitch(image1, "left", False, 0, "white", image2) assert result[0].shape == (1, 32, 56, 3) # 24 + 32 width def test_basic_vertical_stitch_down(self): """Test basic vertical stitching downward""" node = ImageStitch() image1 = self.create_test_image(height=32, width=32) image2 = self.create_test_image(height=24, width=32) result = node.stitch(image1, "down", False, 0, "white", image2) assert result[0].shape == (1, 56, 32, 3) # 32 + 24 height def test_basic_vertical_stitch_up(self): """Test basic vertical stitching upward""" node = ImageStitch() image1 = self.create_test_image(height=32, width=32) image2 = self.create_test_image(height=24, width=32) result = node.stitch(image1, "up", False, 0, "white", image2) assert result[0].shape == (1, 56, 32, 3) # 24 + 32 height def test_size_matching_horizontal(self): """Test size matching for horizontal concatenation""" node = ImageStitch() image1 = self.create_test_image(height=64, width=64) image2 = self.create_test_image(height=32, width=32) # Different aspect ratio result = node.stitch(image1, "right", True, 0, "white", image2) # image2 should be resized to match image1's height (64) with preserved aspect ratio expected_width = 64 + 64 # original + resized (32*64/32 = 64) assert result[0].shape == (1, 64, expected_width, 3) def test_size_matching_vertical(self): """Test size matching for vertical concatenation""" node = ImageStitch() image1 = self.create_test_image(height=64, width=64) image2 = self.create_test_image(height=32, width=32) result = node.stitch(image1, "down", True, 0, "white", image2) # image2 should be resized to match image1's width (64) with preserved aspect ratio expected_height = 64 + 64 # original + resized (32*64/32 = 64) assert result[0].shape == (1, expected_height, 64, 3) def test_padding_for_mismatched_heights_horizontal(self): """Test padding when heights don't match in horizontal concatenation""" node = ImageStitch() image1 = self.create_test_image(height=64, width=32) image2 = self.create_test_image(height=48, width=24) # Shorter height result = node.stitch(image1, "right", False, 0, "white", image2) # Both images should be padded to height 64 assert result[0].shape == (1, 64, 56, 3) # 32 + 24 width, max(64,48) height def test_padding_for_mismatched_widths_vertical(self): """Test padding when widths don't match in vertical concatenation""" node = ImageStitch() image1 = self.create_test_image(height=32, width=64) image2 = self.create_test_image(height=24, width=48) # Narrower width result = node.stitch(image1, "down", False, 0, "white", image2) # Both images should be padded to width 64 assert result[0].shape == (1, 56, 64, 3) # 32 + 24 height, max(64,48) width def test_spacing_horizontal(self): """Test spacing addition in horizontal concatenation""" node = ImageStitch() image1 = self.create_test_image(height=32, width=32) image2 = self.create_test_image(height=32, width=24) spacing_width = 16 result = node.stitch(image1, "right", False, spacing_width, "white", image2) # Expected width: 32 + 16 (spacing) + 24 = 72 assert result[0].shape == (1, 32, 72, 3) def test_spacing_vertical(self): """Test spacing addition in vertical concatenation""" node = ImageStitch() image1 = self.create_test_image(height=32, width=32) image2 = self.create_test_image(height=24, width=32) spacing_width = 16 result = node.stitch(image1, "down", False, spacing_width, "white", image2) # Expected height: 32 + 16 (spacing) + 24 = 72 assert result[0].shape == (1, 72, 32, 3) def test_spacing_color_values(self): """Test that spacing colors are applied correctly""" node = ImageStitch() image1 = self.create_test_image(height=32, width=32) image2 = self.create_test_image(height=32, width=32) # Test white spacing result_white = node.stitch(image1, "right", False, 16, "white", image2) # Check that spacing region contains white values (close to 1.0) spacing_region = result_white[0][:, :, 32:48, :] # Middle 16 pixels assert torch.all(spacing_region >= 0.9) # Should be close to white # Test black spacing result_black = node.stitch(image1, "right", False, 16, "black", image2) spacing_region = result_black[0][:, :, 32:48, :] assert torch.all(spacing_region <= 0.1) # Should be close to black def test_odd_spacing_width_made_even(self): """Test that odd spacing widths are made even""" node = ImageStitch() image1 = self.create_test_image(height=32, width=32) image2 = self.create_test_image(height=32, width=32) # Use odd spacing width result = node.stitch(image1, "right", False, 15, "white", image2) # Should be made even (16), so total width = 32 + 16 + 32 = 80 assert result[0].shape == (1, 32, 80, 3) def test_batch_size_matching(self): """Test that different batch sizes are handled correctly""" node = ImageStitch() image1 = self.create_test_image(batch_size=2, height=32, width=32) image2 = self.create_test_image(batch_size=1, height=32, width=32) result = node.stitch(image1, "right", False, 0, "white", image2) # Should match larger batch size assert result[0].shape == (2, 32, 64, 3) def test_channel_matching_rgb_to_rgba(self): """Test that channel differences are handled (RGB + alpha)""" node = ImageStitch() image1 = self.create_test_image(channels=3) # RGB image2 = self.create_test_image(channels=4) # RGBA result = node.stitch(image1, "right", False, 0, "white", image2) # Should have 4 channels (RGBA) assert result[0].shape[-1] == 4 def test_channel_matching_rgba_to_rgb(self): """Test that channel differences are handled (RGBA + RGB)""" node = ImageStitch() image1 = self.create_test_image(channels=4) # RGBA image2 = self.create_test_image(channels=3) # RGB result = node.stitch(image1, "right", False, 0, "white", image2) # Should have 4 channels (RGBA) assert result[0].shape[-1] == 4 def test_all_color_options(self): """Test all available color options""" node = ImageStitch() image1 = self.create_test_image(height=32, width=32) image2 = self.create_test_image(height=32, width=32) colors = ["white", "black", "red", "green", "blue"] for color in colors: result = node.stitch(image1, "right", False, 16, color, image2) assert result[0].shape == (1, 32, 80, 3) # Basic shape check def test_all_directions(self): """Test all direction options""" node = ImageStitch() image1 = self.create_test_image(height=32, width=32) image2 = self.create_test_image(height=32, width=32) directions = ["right", "left", "up", "down"] for direction in directions: result = node.stitch(image1, direction, False, 0, "white", image2) assert result[0].shape == (1, 32, 64, 3) if direction in ["right", "left"] else (1, 64, 32, 3) def test_batch_size_channel_spacing_integration(self): """Test integration of batch matching, channel matching, size matching, and spacings""" node = ImageStitch() image1 = self.create_test_image(batch_size=2, height=64, width=48, channels=3) image2 = self.create_test_image(batch_size=1, height=32, width=32, channels=4) result = node.stitch(image1, "right", True, 8, "red", image2) # Should handle: batch matching, size matching, channel matching, spacing assert result[0].shape[0] == 2 # Batch size matched assert result[0].shape[-1] == 4 # Channels matched to max assert result[0].shape[1] == 64 # Height from image1 (size matching) # Width should be: 48 + 8 (spacing) + resized_image2_width expected_image2_width = int(64 * (32/32)) # Resized to height 64 expected_total_width = 48 + 8 + expected_image2_width assert result[0].shape[2] == expected_total_width
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "tests-unit/comfy_extras_test/image_stitch_test.py", "license": "GNU General Public License v3.0", "lines": 179, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
Comfy-Org/ComfyUI:comfy_api_nodes/nodes_gemini.py
""" API Nodes for Gemini Multimodal LLM Usage via Remote API See: https://cloud.google.com/vertex-ai/generative-ai/docs/model-reference/inference """ import base64 import os from enum import Enum from fnmatch import fnmatch from io import BytesIO from typing import Literal import torch from typing_extensions import override import folder_paths from comfy_api.latest import IO, ComfyExtension, Input, Types from comfy_api_nodes.apis.gemini import ( GeminiContent, GeminiFileData, GeminiGenerateContentRequest, GeminiGenerateContentResponse, GeminiImageConfig, GeminiImageGenerateContentRequest, GeminiImageGenerationConfig, GeminiInlineData, GeminiMimeType, GeminiPart, GeminiRole, GeminiSystemInstructionContent, GeminiTextPart, GeminiThinkingConfig, Modality, ) from comfy_api_nodes.util import ( ApiEndpoint, audio_to_base64_string, bytesio_to_image_tensor, download_url_to_image_tensor, get_number_of_images, sync_op, tensor_to_base64_string, upload_images_to_comfyapi, validate_string, video_to_base64_string, ) GEMINI_BASE_ENDPOINT = "/proxy/vertexai/gemini" GEMINI_MAX_INPUT_FILE_SIZE = 20 * 1024 * 1024 # 20 MB GEMINI_IMAGE_SYS_PROMPT = ( "You are an expert image-generation engine. You must ALWAYS produce an image.\n" "Interpret all user input—regardless of " "format, intent, or abstraction—as literal visual directives for image composition.\n" "If a prompt is conversational or lacks specific visual details, " "you must creatively invent a concrete visual scenario that depicts the concept.\n" "Prioritize generating the visual representation above any text, formatting, or conversational requests." ) GEMINI_IMAGE_2_PRICE_BADGE = IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["model", "resolution"]), expr=""" ( $m := widgets.model; $r := widgets.resolution; $isFlash := $contains($m, "nano banana 2"); $flashPrices := {"1k": 0.0696, "2k": 0.0696, "4k": 0.123}; $proPrices := {"1k": 0.134, "2k": 0.134, "4k": 0.24}; $prices := $isFlash ? $flashPrices : $proPrices; {"type":"usd","usd": $lookup($prices, $r), "format":{"suffix":"/Image","approximate":true}} ) """, ) class GeminiModel(str, Enum): """ Gemini Model Names allowed by comfy-api """ gemini_2_5_pro_preview_05_06 = "gemini-2.5-pro-preview-05-06" gemini_2_5_flash_preview_04_17 = "gemini-2.5-flash-preview-04-17" gemini_2_5_pro = "gemini-2.5-pro" gemini_2_5_flash = "gemini-2.5-flash" gemini_3_0_pro = "gemini-3-pro-preview" class GeminiImageModel(str, Enum): """ Gemini Image Model Names allowed by comfy-api """ gemini_2_5_flash_image_preview = "gemini-2.5-flash-image-preview" gemini_2_5_flash_image = "gemini-2.5-flash-image" async def create_image_parts( cls: type[IO.ComfyNode], images: Input.Image, image_limit: int = 0, ) -> list[GeminiPart]: image_parts: list[GeminiPart] = [] if image_limit < 0: raise ValueError("image_limit must be greater than or equal to 0 when creating Gemini image parts.") total_images = get_number_of_images(images) if total_images <= 0: raise ValueError("No images provided to create_image_parts; at least one image is required.") # If image_limit == 0 --> use all images; otherwise clamp to image_limit. effective_max = total_images if image_limit == 0 else min(total_images, image_limit) # Number of images we'll send as URLs (fileData) num_url_images = min(effective_max, 10) # Vertex API max number of image links reference_images_urls = await upload_images_to_comfyapi( cls, images, max_images=num_url_images, ) for reference_image_url in reference_images_urls: image_parts.append( GeminiPart( fileData=GeminiFileData( mimeType=GeminiMimeType.image_png, fileUri=reference_image_url, ) ) ) for idx in range(num_url_images, effective_max): image_parts.append( GeminiPart( inlineData=GeminiInlineData( mimeType=GeminiMimeType.image_png, data=tensor_to_base64_string(images[idx]), ) ) ) return image_parts def _mime_matches(mime: GeminiMimeType | None, pattern: str) -> bool: """Check if a MIME type matches a pattern. Supports fnmatch globs (e.g. 'image/*').""" if mime is None: return False return fnmatch(mime.value, pattern) def get_parts_by_type(response: GeminiGenerateContentResponse, part_type: Literal["text"] | str) -> list[GeminiPart]: """ Filter response parts by their type. Args: response: The API response from Gemini. part_type: Type of parts to extract ("text" or a MIME type). Returns: List of response parts matching the requested type. """ if not response.candidates: if response.promptFeedback and response.promptFeedback.blockReason: feedback = response.promptFeedback raise ValueError( f"Gemini API blocked the request. Reason: {feedback.blockReason} ({feedback.blockReasonMessage})" ) raise ValueError( "Gemini API returned no response candidates. If you are using the `IMAGE` modality, " "try changing it to `IMAGE+TEXT` to view the model's reasoning and understand why image generation failed." ) parts = [] blocked_reasons = [] for candidate in response.candidates: if candidate.finishReason and candidate.finishReason.upper() == "IMAGE_PROHIBITED_CONTENT": blocked_reasons.append(candidate.finishReason) continue if candidate.content is None or candidate.content.parts is None: continue for part in candidate.content.parts: if part_type == "text" and part.text: parts.append(part) elif part.inlineData and _mime_matches(part.inlineData.mimeType, part_type): parts.append(part) elif part.fileData and _mime_matches(part.fileData.mimeType, part_type): parts.append(part) if not parts and blocked_reasons: raise ValueError(f"Gemini API blocked the request. Reasons: {blocked_reasons}") return parts def get_text_from_response(response: GeminiGenerateContentResponse) -> str: """ Extract and concatenate all text parts from the response. Args: response: The API response from Gemini. Returns: Combined text from all text parts in the response. """ parts = get_parts_by_type(response, "text") return "\n".join([part.text for part in parts]) async def get_image_from_response(response: GeminiGenerateContentResponse) -> Input.Image: image_tensors: list[Input.Image] = [] parts = get_parts_by_type(response, "image/*") for part in parts: if part.inlineData: image_data = base64.b64decode(part.inlineData.data) returned_image = bytesio_to_image_tensor(BytesIO(image_data)) else: returned_image = await download_url_to_image_tensor(part.fileData.fileUri) image_tensors.append(returned_image) if len(image_tensors) == 0: return torch.zeros((1, 1024, 1024, 4)) return torch.cat(image_tensors, dim=0) def calculate_tokens_price(response: GeminiGenerateContentResponse) -> float | None: if not response.modelVersion: return None # Define prices (Cost per 1,000,000 tokens), see https://cloud.google.com/vertex-ai/generative-ai/pricing if response.modelVersion in ("gemini-2.5-pro-preview-05-06", "gemini-2.5-pro"): input_tokens_price = 1.25 output_text_tokens_price = 10.0 output_image_tokens_price = 0.0 elif response.modelVersion in ( "gemini-2.5-flash-preview-04-17", "gemini-2.5-flash", ): input_tokens_price = 0.30 output_text_tokens_price = 2.50 output_image_tokens_price = 0.0 elif response.modelVersion in ( "gemini-2.5-flash-image-preview", "gemini-2.5-flash-image", ): input_tokens_price = 0.30 output_text_tokens_price = 2.50 output_image_tokens_price = 30.0 elif response.modelVersion == "gemini-3-pro-preview": input_tokens_price = 2 output_text_tokens_price = 12.0 output_image_tokens_price = 0.0 elif response.modelVersion == "gemini-3-pro-image-preview": input_tokens_price = 2 output_text_tokens_price = 12.0 output_image_tokens_price = 120.0 elif response.modelVersion == "gemini-3.1-flash-image-preview": input_tokens_price = 0.5 output_text_tokens_price = 3.0 output_image_tokens_price = 60.0 else: return None final_price = response.usageMetadata.promptTokenCount * input_tokens_price if response.usageMetadata.candidatesTokensDetails: for i in response.usageMetadata.candidatesTokensDetails: if i.modality == Modality.IMAGE: final_price += output_image_tokens_price * i.tokenCount # for Nano Banana models else: final_price += output_text_tokens_price * i.tokenCount if response.usageMetadata.thoughtsTokenCount: final_price += output_text_tokens_price * response.usageMetadata.thoughtsTokenCount return final_price / 1_000_000.0 class GeminiNode(IO.ComfyNode): """ Node to generate text responses from a Gemini model. This node allows users to interact with Google's Gemini AI models, providing multimodal inputs (text, images, audio, video, files) to generate coherent text responses. The node works with the latest Gemini models, handling the API communication and response parsing. """ @classmethod def define_schema(cls): return IO.Schema( node_id="GeminiNode", display_name="Google Gemini", category="api node/text/Gemini", description="Generate text responses with Google's Gemini AI model. " "You can provide multiple types of inputs (text, images, audio, video) " "as context for generating more relevant and meaningful responses.", inputs=[ IO.String.Input( "prompt", multiline=True, default="", tooltip="Text inputs to the model, used to generate a response. " "You can include detailed instructions, questions, or context for the model.", ), IO.Combo.Input( "model", options=GeminiModel, default=GeminiModel.gemini_2_5_pro, tooltip="The Gemini model to use for generating responses.", ), IO.Int.Input( "seed", default=42, min=0, max=0xFFFFFFFFFFFFFFFF, control_after_generate=True, tooltip="When seed is fixed to a specific value, the model makes a best effort to provide " "the same response for repeated requests. Deterministic output isn't guaranteed. " "Also, changing the model or parameter settings, such as the temperature, " "can cause variations in the response even when you use the same seed value. " "By default, a random seed value is used.", ), IO.Image.Input( "images", optional=True, tooltip="Optional image(s) to use as context for the model. " "To include multiple images, you can use the Batch Images node.", ), IO.Audio.Input( "audio", optional=True, tooltip="Optional audio to use as context for the model.", ), IO.Video.Input( "video", optional=True, tooltip="Optional video to use as context for the model.", ), IO.Custom("GEMINI_INPUT_FILES").Input( "files", optional=True, tooltip="Optional file(s) to use as context for the model. " "Accepts inputs from the Gemini Generate Content Input Files node.", ), IO.String.Input( "system_prompt", multiline=True, default="", optional=True, tooltip="Foundational instructions that dictate an AI's behavior.", advanced=True, ), ], outputs=[ IO.String.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["model"]), expr=""" ( $m := widgets.model; $contains($m, "gemini-2.5-flash") ? { "type": "list_usd", "usd": [0.0003, 0.0025], "format": { "approximate": true, "separator": "-", "suffix": " per 1K tokens"} } : $contains($m, "gemini-2.5-pro") ? { "type": "list_usd", "usd": [0.00125, 0.01], "format": { "approximate": true, "separator": "-", "suffix": " per 1K tokens" } } : $contains($m, "gemini-3-pro-preview") ? { "type": "list_usd", "usd": [0.002, 0.012], "format": { "approximate": true, "separator": "-", "suffix": " per 1K tokens" } } : {"type":"text", "text":"Token-based"} ) """, ), ) @classmethod def create_video_parts(cls, video_input: Input.Video) -> list[GeminiPart]: """Convert video input to Gemini API compatible parts.""" base_64_string = video_to_base64_string( video_input, container_format=Types.VideoContainer.MP4, codec=Types.VideoCodec.H264 ) return [ GeminiPart( inlineData=GeminiInlineData( mimeType=GeminiMimeType.video_mp4, data=base_64_string, ) ) ] @classmethod def create_audio_parts(cls, audio_input: Input.Audio) -> list[GeminiPart]: """ Convert audio input to Gemini API compatible parts. Args: audio_input: Audio input from ComfyUI, containing waveform tensor and sample rate. Returns: List of GeminiPart objects containing the encoded audio. """ audio_parts: list[GeminiPart] = [] for batch_index in range(audio_input["waveform"].shape[0]): # Recreate an IO.AUDIO object for the given batch dimension index audio_at_index = Input.Audio( waveform=audio_input["waveform"][batch_index].unsqueeze(0), sample_rate=audio_input["sample_rate"], ) # Convert to MP3 format for compatibility with Gemini API audio_bytes = audio_to_base64_string( audio_at_index, container_format="mp3", codec_name="libmp3lame", ) audio_parts.append( GeminiPart( inlineData=GeminiInlineData( mimeType=GeminiMimeType.audio_mp3, data=audio_bytes, ) ) ) return audio_parts @classmethod async def execute( cls, prompt: str, model: str, seed: int, images: Input.Image | None = None, audio: Input.Audio | None = None, video: Input.Video | None = None, files: list[GeminiPart] | None = None, system_prompt: str = "", ) -> IO.NodeOutput: validate_string(prompt, strip_whitespace=False) # Create parts list with text prompt as the first part parts: list[GeminiPart] = [GeminiPart(text=prompt)] # Add other modal parts if images is not None: parts.extend(await create_image_parts(cls, images)) if audio is not None: parts.extend(cls.create_audio_parts(audio)) if video is not None: parts.extend(cls.create_video_parts(video)) if files is not None: parts.extend(files) gemini_system_prompt = None if system_prompt: gemini_system_prompt = GeminiSystemInstructionContent(parts=[GeminiTextPart(text=system_prompt)], role=None) response = await sync_op( cls, endpoint=ApiEndpoint(path=f"{GEMINI_BASE_ENDPOINT}/{model}", method="POST"), data=GeminiGenerateContentRequest( contents=[ GeminiContent( role=GeminiRole.user, parts=parts, ) ], systemInstruction=gemini_system_prompt, ), response_model=GeminiGenerateContentResponse, price_extractor=calculate_tokens_price, ) output_text = get_text_from_response(response) return IO.NodeOutput(output_text or "Empty response from Gemini model...") class GeminiInputFiles(IO.ComfyNode): """ Loads and formats input files for use with the Gemini API. This node allows users to include text (.txt) and PDF (.pdf) files as input context for the Gemini model. Files are converted to the appropriate format required by the API and can be chained together to include multiple files in a single request. """ @classmethod def define_schema(cls): """ For details about the supported file input types, see: https://cloud.google.com/vertex-ai/generative-ai/docs/model-reference/inference """ input_dir = folder_paths.get_input_directory() input_files = [ f for f in os.scandir(input_dir) if f.is_file() and (f.name.endswith(".txt") or f.name.endswith(".pdf")) and f.stat().st_size < GEMINI_MAX_INPUT_FILE_SIZE ] input_files = sorted(input_files, key=lambda x: x.name) input_files = [f.name for f in input_files] return IO.Schema( node_id="GeminiInputFiles", display_name="Gemini Input Files", category="api node/text/Gemini", description="Loads and prepares input files to include as inputs for Gemini LLM nodes. " "The files will be read by the Gemini model when generating a response. " "The contents of the text file count toward the token limit. " "🛈 TIP: Can be chained together with other Gemini Input File nodes.", inputs=[ IO.Combo.Input( "file", options=input_files, default=input_files[0] if input_files else None, tooltip="Input files to include as context for the model. " "Only accepts text (.txt) and PDF (.pdf) files for now.", ), IO.Custom("GEMINI_INPUT_FILES").Input( "GEMINI_INPUT_FILES", optional=True, tooltip="An optional additional file(s) to batch together with the file loaded from this node. " "Allows chaining of input files so that a single message can include multiple input files.", ), ], outputs=[ IO.Custom("GEMINI_INPUT_FILES").Output(), ], ) @classmethod def create_file_part(cls, file_path: str) -> GeminiPart: mime_type = GeminiMimeType.application_pdf if file_path.endswith(".pdf") else GeminiMimeType.text_plain # Use base64 string directly, not the data URI with open(file_path, "rb") as f: file_content = f.read() base64_str = base64.b64encode(file_content).decode("utf-8") return GeminiPart( inlineData=GeminiInlineData( mimeType=mime_type, data=base64_str, ) ) @classmethod def execute(cls, file: str, GEMINI_INPUT_FILES: list[GeminiPart] | None = None) -> IO.NodeOutput: """Loads and formats input files for Gemini API.""" if GEMINI_INPUT_FILES is None: GEMINI_INPUT_FILES = [] file_path = folder_paths.get_annotated_filepath(file) input_file_content = cls.create_file_part(file_path) return IO.NodeOutput([input_file_content] + GEMINI_INPUT_FILES) class GeminiImage(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="GeminiImageNode", display_name="Nano Banana (Google Gemini Image)", category="api node/image/Gemini", description="Edit images synchronously via Google API.", inputs=[ IO.String.Input( "prompt", multiline=True, tooltip="Text prompt for generation", default="", ), IO.Combo.Input( "model", options=GeminiImageModel, default=GeminiImageModel.gemini_2_5_flash_image, tooltip="The Gemini model to use for generating responses.", ), IO.Int.Input( "seed", default=42, min=0, max=0xFFFFFFFFFFFFFFFF, control_after_generate=True, tooltip="When seed is fixed to a specific value, the model makes a best effort to provide " "the same response for repeated requests. Deterministic output isn't guaranteed. " "Also, changing the model or parameter settings, such as the temperature, " "can cause variations in the response even when you use the same seed value. " "By default, a random seed value is used.", ), IO.Image.Input( "images", optional=True, tooltip="Optional image(s) to use as context for the model. " "To include multiple images, you can use the Batch Images node.", ), IO.Custom("GEMINI_INPUT_FILES").Input( "files", optional=True, tooltip="Optional file(s) to use as context for the model. " "Accepts inputs from the Gemini Generate Content Input Files node.", ), IO.Combo.Input( "aspect_ratio", options=["auto", "1:1", "2:3", "3:2", "3:4", "4:3", "4:5", "5:4", "9:16", "16:9", "21:9"], default="auto", tooltip="Defaults to matching the output image size to that of your input image, " "or otherwise generates 1:1 squares.", optional=True, ), IO.Combo.Input( "response_modalities", options=["IMAGE+TEXT", "IMAGE"], tooltip="Choose 'IMAGE' for image-only output, or " "'IMAGE+TEXT' to return both the generated image and a text response.", optional=True, advanced=True, ), IO.String.Input( "system_prompt", multiline=True, default=GEMINI_IMAGE_SYS_PROMPT, optional=True, tooltip="Foundational instructions that dictate an AI's behavior.", advanced=True, ), ], outputs=[ IO.Image.Output(), IO.String.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.039,"format":{"suffix":"/Image (1K)","approximate":true}}""", ), ) @classmethod async def execute( cls, prompt: str, model: str, seed: int, images: Input.Image | None = None, files: list[GeminiPart] | None = None, aspect_ratio: str = "auto", response_modalities: str = "IMAGE+TEXT", system_prompt: str = "", ) -> IO.NodeOutput: validate_string(prompt, strip_whitespace=True, min_length=1) parts: list[GeminiPart] = [GeminiPart(text=prompt)] if not aspect_ratio: aspect_ratio = "auto" # for backward compatability with old workflows; to-do remove this in December image_config = GeminiImageConfig() if aspect_ratio == "auto" else GeminiImageConfig(aspectRatio=aspect_ratio) if images is not None: parts.extend(await create_image_parts(cls, images)) if files is not None: parts.extend(files) gemini_system_prompt = None if system_prompt: gemini_system_prompt = GeminiSystemInstructionContent(parts=[GeminiTextPart(text=system_prompt)], role=None) response = await sync_op( cls, ApiEndpoint(path=f"/proxy/vertexai/gemini/{model}", method="POST"), data=GeminiImageGenerateContentRequest( contents=[ GeminiContent(role=GeminiRole.user, parts=parts), ], generationConfig=GeminiImageGenerationConfig( responseModalities=(["IMAGE"] if response_modalities == "IMAGE" else ["TEXT", "IMAGE"]), imageConfig=image_config, ), systemInstruction=gemini_system_prompt, ), response_model=GeminiGenerateContentResponse, price_extractor=calculate_tokens_price, ) return IO.NodeOutput(await get_image_from_response(response), get_text_from_response(response)) class GeminiImage2(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="GeminiImage2Node", display_name="Nano Banana Pro (Google Gemini Image)", category="api node/image/Gemini", description="Generate or edit images synchronously via Google Vertex API.", inputs=[ IO.String.Input( "prompt", multiline=True, tooltip="Text prompt describing the image to generate or the edits to apply. " "Include any constraints, styles, or details the model should follow.", default="", ), IO.Combo.Input( "model", options=["gemini-3-pro-image-preview", "Nano Banana 2 (Gemini 3.1 Flash Image)"], ), IO.Int.Input( "seed", default=42, min=0, max=0xFFFFFFFFFFFFFFFF, control_after_generate=True, tooltip="When the seed is fixed to a specific value, the model makes a best effort to provide " "the same response for repeated requests. Deterministic output isn't guaranteed. " "Also, changing the model or parameter settings, such as the temperature, " "can cause variations in the response even when you use the same seed value. " "By default, a random seed value is used.", ), IO.Combo.Input( "aspect_ratio", options=["auto", "1:1", "2:3", "3:2", "3:4", "4:3", "4:5", "5:4", "9:16", "16:9", "21:9"], default="auto", tooltip="If set to 'auto', matches your input image's aspect ratio; " "if no image is provided, a 16:9 square is usually generated.", ), IO.Combo.Input( "resolution", options=["1K", "2K", "4K"], tooltip="Target output resolution. For 2K/4K the native Gemini upscaler is used.", ), IO.Combo.Input( "response_modalities", options=["IMAGE+TEXT", "IMAGE"], tooltip="Choose 'IMAGE' for image-only output, or " "'IMAGE+TEXT' to return both the generated image and a text response.", advanced=True, ), IO.Image.Input( "images", optional=True, tooltip="Optional reference image(s). " "To include multiple images, use the Batch Images node (up to 14).", ), IO.Custom("GEMINI_INPUT_FILES").Input( "files", optional=True, tooltip="Optional file(s) to use as context for the model. " "Accepts inputs from the Gemini Generate Content Input Files node.", ), IO.String.Input( "system_prompt", multiline=True, default=GEMINI_IMAGE_SYS_PROMPT, optional=True, tooltip="Foundational instructions that dictate an AI's behavior.", advanced=True, ), ], outputs=[ IO.Image.Output(), IO.String.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=GEMINI_IMAGE_2_PRICE_BADGE, ) @classmethod async def execute( cls, prompt: str, model: str, seed: int, aspect_ratio: str, resolution: str, response_modalities: str, images: Input.Image | None = None, files: list[GeminiPart] | None = None, system_prompt: str = "", ) -> IO.NodeOutput: validate_string(prompt, strip_whitespace=True, min_length=1) if model == "Nano Banana 2 (Gemini 3.1 Flash Image)": model = "gemini-3.1-flash-image-preview" if response_modalities == "IMAGE+TEXT": raise ValueError("IMAGE+TEXT is not currently available for the Nano Banana 2 model.") parts: list[GeminiPart] = [GeminiPart(text=prompt)] if images is not None: if get_number_of_images(images) > 14: raise ValueError("The current maximum number of supported images is 14.") parts.extend(await create_image_parts(cls, images)) if files is not None: parts.extend(files) image_config = GeminiImageConfig(imageSize=resolution) if aspect_ratio != "auto": image_config.aspectRatio = aspect_ratio gemini_system_prompt = None if system_prompt: gemini_system_prompt = GeminiSystemInstructionContent(parts=[GeminiTextPart(text=system_prompt)], role=None) response = await sync_op( cls, ApiEndpoint(path=f"/proxy/vertexai/gemini/{model}", method="POST"), data=GeminiImageGenerateContentRequest( contents=[ GeminiContent(role=GeminiRole.user, parts=parts), ], generationConfig=GeminiImageGenerationConfig( responseModalities=(["IMAGE"] if response_modalities == "IMAGE" else ["TEXT", "IMAGE"]), imageConfig=image_config, ), systemInstruction=gemini_system_prompt, ), response_model=GeminiGenerateContentResponse, price_extractor=calculate_tokens_price, ) return IO.NodeOutput(await get_image_from_response(response), get_text_from_response(response)) class GeminiNanoBanana2(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="GeminiNanoBanana2", display_name="Nano Banana 2", category="api node/image/Gemini", description="Generate or edit images synchronously via Google Vertex API.", inputs=[ IO.String.Input( "prompt", multiline=True, tooltip="Text prompt describing the image to generate or the edits to apply. " "Include any constraints, styles, or details the model should follow.", default="", ), IO.Combo.Input( "model", options=["Nano Banana 2 (Gemini 3.1 Flash Image)"], ), IO.Int.Input( "seed", default=42, min=0, max=0xFFFFFFFFFFFFFFFF, control_after_generate=True, tooltip="When the seed is fixed to a specific value, the model makes a best effort to provide " "the same response for repeated requests. Deterministic output isn't guaranteed. " "Also, changing the model or parameter settings, such as the temperature, " "can cause variations in the response even when you use the same seed value. " "By default, a random seed value is used.", ), IO.Combo.Input( "aspect_ratio", options=[ "auto", "1:1", "2:3", "3:2", "3:4", "4:3", "4:5", "5:4", "9:16", "16:9", "21:9", # "1:4", # "4:1", # "8:1", # "1:8", ], default="auto", tooltip="If set to 'auto', matches your input image's aspect ratio; " "if no image is provided, a 16:9 square is usually generated.", ), IO.Combo.Input( "resolution", options=[ # "512px", "1K", "2K", "4K", ], tooltip="Target output resolution. For 2K/4K the native Gemini upscaler is used.", ), IO.Combo.Input( "response_modalities", options=["IMAGE"], advanced=True, ), IO.Combo.Input( "thinking_level", options=["MINIMAL", "HIGH"], ), IO.Image.Input( "images", optional=True, tooltip="Optional reference image(s). " "To include multiple images, use the Batch Images node (up to 14).", ), IO.Custom("GEMINI_INPUT_FILES").Input( "files", optional=True, tooltip="Optional file(s) to use as context for the model. " "Accepts inputs from the Gemini Generate Content Input Files node.", ), IO.String.Input( "system_prompt", multiline=True, default=GEMINI_IMAGE_SYS_PROMPT, optional=True, tooltip="Foundational instructions that dictate an AI's behavior.", advanced=True, ), ], outputs=[ IO.Image.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=GEMINI_IMAGE_2_PRICE_BADGE, ) @classmethod async def execute( cls, prompt: str, model: str, seed: int, aspect_ratio: str, resolution: str, response_modalities: str, thinking_level: str, images: Input.Image | None = None, files: list[GeminiPart] | None = None, system_prompt: str = "", ) -> IO.NodeOutput: validate_string(prompt, strip_whitespace=True, min_length=1) if model == "Nano Banana 2 (Gemini 3.1 Flash Image)": model = "gemini-3.1-flash-image-preview" parts: list[GeminiPart] = [GeminiPart(text=prompt)] if images is not None: if get_number_of_images(images) > 14: raise ValueError("The current maximum number of supported images is 14.") parts.extend(await create_image_parts(cls, images)) if files is not None: parts.extend(files) image_config = GeminiImageConfig(imageSize=resolution) if aspect_ratio != "auto": image_config.aspectRatio = aspect_ratio gemini_system_prompt = None if system_prompt: gemini_system_prompt = GeminiSystemInstructionContent(parts=[GeminiTextPart(text=system_prompt)], role=None) response = await sync_op( cls, ApiEndpoint(path=f"/proxy/vertexai/gemini/{model}", method="POST"), data=GeminiImageGenerateContentRequest( contents=[ GeminiContent(role=GeminiRole.user, parts=parts), ], generationConfig=GeminiImageGenerationConfig( responseModalities=(["IMAGE"] if response_modalities == "IMAGE" else ["TEXT", "IMAGE"]), imageConfig=image_config, thinkingConfig=GeminiThinkingConfig(thinkingLevel=thinking_level), ), systemInstruction=gemini_system_prompt, ), response_model=GeminiGenerateContentResponse, price_extractor=calculate_tokens_price, ) return IO.NodeOutput(await get_image_from_response(response), get_text_from_response(response)) class GeminiExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[IO.ComfyNode]]: return [ GeminiNode, GeminiImage, GeminiImage2, GeminiNanoBanana2, GeminiInputFiles, ] async def comfy_entrypoint() -> GeminiExtension: return GeminiExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api_nodes/nodes_gemini.py", "license": "GNU General Public License v3.0", "lines": 924, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_api_nodes/nodes_rodin.py
""" ComfyUI X Rodin3D(Deemos) API Nodes Rodin API docs: https://developer.hyper3d.ai/ """ from inspect import cleandoc import folder_paths as comfy_paths import os import logging import math from io import BytesIO from typing_extensions import override from PIL import Image from comfy_api_nodes.apis.rodin import ( Rodin3DGenerateRequest, Rodin3DGenerateResponse, Rodin3DCheckStatusRequest, Rodin3DCheckStatusResponse, Rodin3DDownloadRequest, Rodin3DDownloadResponse, JobStatus, ) from comfy_api_nodes.util import ( sync_op, poll_op, ApiEndpoint, download_url_to_bytesio, download_url_to_file_3d, ) from comfy_api.latest import ComfyExtension, IO, Types COMMON_PARAMETERS = [ IO.Int.Input( "Seed", default=0, min=0, max=65535, display_mode=IO.NumberDisplay.number, optional=True, ), IO.Combo.Input("Material_Type", options=["PBR", "Shaded"], default="PBR", optional=True), IO.Combo.Input( "Polygon_count", options=["4K-Quad", "8K-Quad", "18K-Quad", "50K-Quad", "200K-Triangle"], default="18K-Quad", optional=True, ), ] def get_quality_mode(poly_count): polycount = poly_count.split("-") poly = polycount[1] count = polycount[0] if poly == "Triangle": mesh_mode = "Raw" elif poly == "Quad": mesh_mode = "Quad" else: mesh_mode = "Quad" if count == "4K": quality_override = 4000 elif count == "8K": quality_override = 8000 elif count == "18K": quality_override = 18000 elif count == "50K": quality_override = 50000 elif count == "2K": quality_override = 2000 elif count == "20K": quality_override = 20000 elif count == "150K": quality_override = 150000 elif count == "500K": quality_override = 500000 else: quality_override = 18000 return mesh_mode, quality_override def tensor_to_filelike(tensor, max_pixels: int = 2048*2048): """ Converts a PyTorch tensor to a file-like object. Args: - tensor (torch.Tensor): A tensor representing an image of shape (H, W, C) where C is the number of channels (3 for RGB), H is height, and W is width. Returns: - io.BytesIO: A file-like object containing the image data. """ array = tensor.cpu().numpy() array = (array * 255).astype('uint8') image = Image.fromarray(array, 'RGB') original_width, original_height = image.size original_pixels = original_width * original_height if original_pixels > max_pixels: scale = math.sqrt(max_pixels / original_pixels) new_width = int(original_width * scale) new_height = int(original_height * scale) else: new_width, new_height = original_width, original_height if new_width != original_width or new_height != original_height: image = image.resize((new_width, new_height), Image.Resampling.LANCZOS) img_byte_arr = BytesIO() image.save(img_byte_arr, format='PNG') # PNG is used for lossless compression img_byte_arr.seek(0) return img_byte_arr async def create_generate_task( cls: type[IO.ComfyNode], images=None, seed=1, material="PBR", quality_override=18000, tier="Regular", mesh_mode="Quad", ta_pose: bool = False, ): if images is None: raise Exception("Rodin 3D generate requires at least 1 image.") if len(images) > 5: raise Exception("Rodin 3D generate requires up to 5 image.") response = await sync_op( cls, ApiEndpoint(path="/proxy/rodin/api/v2/rodin", method="POST"), response_model=Rodin3DGenerateResponse, data=Rodin3DGenerateRequest( seed=seed, tier=tier, material=material, quality_override=quality_override, mesh_mode=mesh_mode, TAPose=ta_pose, ), files=[ ( "images", open(image, "rb") if isinstance(image, str) else tensor_to_filelike(image) ) for image in images if image is not None ], content_type="multipart/form-data", ) if hasattr(response, "error"): error_message = f"Rodin3D Create 3D generate Task Failed. Message: {response.message}, error: {response.error}" logging.error(error_message) raise Exception(error_message) logging.info("[ Rodin3D API - Submit Jobs ] Submit Generate Task Success!") subscription_key = response.jobs.subscription_key task_uuid = response.uuid logging.info("[ Rodin3D API - Submit Jobs ] UUID: %s", task_uuid) return task_uuid, subscription_key def check_rodin_status(response: Rodin3DCheckStatusResponse) -> str: all_done = all(job.status == JobStatus.Done for job in response.jobs) status_list = [str(job.status) for job in response.jobs] logging.info("[ Rodin3D API - CheckStatus ] Generate Status: %s", status_list) if any(job.status == JobStatus.Failed for job in response.jobs): logging.error("[ Rodin3D API - CheckStatus ] Generate Failed: %s, Please try again.", status_list) raise Exception("[ Rodin3D API ] Generate Failed, Please Try again.") if all_done: return "DONE" return "Generating" def extract_progress(response: Rodin3DCheckStatusResponse) -> int | None: if not response.jobs: return None completed_count = sum(1 for job in response.jobs if job.status == JobStatus.Done) return int((completed_count / len(response.jobs)) * 100) async def poll_for_task_status(subscription_key: str, cls: type[IO.ComfyNode]) -> Rodin3DCheckStatusResponse: logging.info("[ Rodin3D API - CheckStatus ] Generate Start!") return await poll_op( cls, ApiEndpoint(path="/proxy/rodin/api/v2/status", method="POST"), response_model=Rodin3DCheckStatusResponse, data=Rodin3DCheckStatusRequest(subscription_key=subscription_key), status_extractor=check_rodin_status, progress_extractor=extract_progress, ) async def get_rodin_download_list(uuid: str, cls: type[IO.ComfyNode]) -> Rodin3DDownloadResponse: logging.info("[ Rodin3D API - Downloading ] Generate Successfully!") return await sync_op( cls, ApiEndpoint(path="/proxy/rodin/api/v2/download", method="POST"), response_model=Rodin3DDownloadResponse, data=Rodin3DDownloadRequest(task_uuid=uuid), monitor_progress=False, ) async def download_files(url_list, task_uuid: str) -> tuple[str | None, Types.File3D | None]: result_folder_name = f"Rodin3D_{task_uuid}" save_path = os.path.join(comfy_paths.get_output_directory(), result_folder_name) os.makedirs(save_path, exist_ok=True) model_file_path = None file_3d = None for i in url_list.list: file_path = os.path.join(save_path, i.name) if i.name.lower().endswith(".glb"): model_file_path = os.path.join(result_folder_name, i.name) file_3d = await download_url_to_file_3d(i.url, "glb") # Save to disk for backward compatibility with open(file_path, "wb") as f: f.write(file_3d.get_bytes()) else: await download_url_to_bytesio(i.url, file_path) return model_file_path, file_3d class Rodin3D_Regular(IO.ComfyNode): """Generate 3D Assets using Rodin API""" @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id="Rodin3D_Regular", display_name="Rodin 3D Generate - Regular Generate", category="api node/3d/Rodin", description=cleandoc(cls.__doc__ or ""), inputs=[ IO.Image.Input("Images"), *COMMON_PARAMETERS, ], outputs=[ IO.String.Output(display_name="3D Model Path"), # for backward compatibility only IO.File3DGLB.Output(display_name="GLB"), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.4}""", ), ) @classmethod async def execute( cls, Images, Seed, Material_Type, Polygon_count, ) -> IO.NodeOutput: tier = "Regular" num_images = Images.shape[0] m_images = [] for i in range(num_images): m_images.append(Images[i]) mesh_mode, quality_override = get_quality_mode(Polygon_count) task_uuid, subscription_key = await create_generate_task( cls, images=m_images, seed=Seed, material=Material_Type, quality_override=quality_override, tier=tier, mesh_mode=mesh_mode, ) await poll_for_task_status(subscription_key, cls) download_list = await get_rodin_download_list(task_uuid, cls) model_path, file_3d = await download_files(download_list, task_uuid) return IO.NodeOutput(model_path, file_3d) class Rodin3D_Detail(IO.ComfyNode): """Generate 3D Assets using Rodin API""" @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id="Rodin3D_Detail", display_name="Rodin 3D Generate - Detail Generate", category="api node/3d/Rodin", description=cleandoc(cls.__doc__ or ""), inputs=[ IO.Image.Input("Images"), *COMMON_PARAMETERS, ], outputs=[ IO.String.Output(display_name="3D Model Path"), # for backward compatibility only IO.File3DGLB.Output(display_name="GLB"), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.4}""", ), ) @classmethod async def execute( cls, Images, Seed, Material_Type, Polygon_count, ) -> IO.NodeOutput: tier = "Detail" num_images = Images.shape[0] m_images = [] for i in range(num_images): m_images.append(Images[i]) mesh_mode, quality_override = get_quality_mode(Polygon_count) task_uuid, subscription_key = await create_generate_task( cls, images=m_images, seed=Seed, material=Material_Type, quality_override=quality_override, tier=tier, mesh_mode=mesh_mode, ) await poll_for_task_status(subscription_key, cls) download_list = await get_rodin_download_list(task_uuid, cls) model_path, file_3d = await download_files(download_list, task_uuid) return IO.NodeOutput(model_path, file_3d) class Rodin3D_Smooth(IO.ComfyNode): """Generate 3D Assets using Rodin API""" @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id="Rodin3D_Smooth", display_name="Rodin 3D Generate - Smooth Generate", category="api node/3d/Rodin", description=cleandoc(cls.__doc__ or ""), inputs=[ IO.Image.Input("Images"), *COMMON_PARAMETERS, ], outputs=[ IO.String.Output(display_name="3D Model Path"), # for backward compatibility only IO.File3DGLB.Output(display_name="GLB"), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.4}""", ), ) @classmethod async def execute( cls, Images, Seed, Material_Type, Polygon_count, ) -> IO.NodeOutput: num_images = Images.shape[0] m_images = [] for i in range(num_images): m_images.append(Images[i]) mesh_mode, quality_override = get_quality_mode(Polygon_count) task_uuid, subscription_key = await create_generate_task( cls, images=m_images, seed=Seed, material=Material_Type, quality_override=quality_override, tier="Smooth", mesh_mode=mesh_mode, ) await poll_for_task_status(subscription_key, cls) download_list = await get_rodin_download_list(task_uuid, cls) model_path, file_3d = await download_files(download_list, task_uuid) return IO.NodeOutput(model_path, file_3d) class Rodin3D_Sketch(IO.ComfyNode): """Generate 3D Assets using Rodin API""" @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id="Rodin3D_Sketch", display_name="Rodin 3D Generate - Sketch Generate", category="api node/3d/Rodin", description=cleandoc(cls.__doc__ or ""), inputs=[ IO.Image.Input("Images"), IO.Int.Input( "Seed", default=0, min=0, max=65535, display_mode=IO.NumberDisplay.number, optional=True, ), ], outputs=[ IO.String.Output(display_name="3D Model Path"), # for backward compatibility only IO.File3DGLB.Output(display_name="GLB"), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.4}""", ), ) @classmethod async def execute( cls, Images, Seed, ) -> IO.NodeOutput: num_images = Images.shape[0] m_images = [] for i in range(num_images): m_images.append(Images[i]) task_uuid, subscription_key = await create_generate_task( cls, images=m_images, seed=Seed, material="PBR", quality_override=18000, tier="Sketch", mesh_mode="Quad", ) await poll_for_task_status(subscription_key, cls) download_list = await get_rodin_download_list(task_uuid, cls) model_path, file_3d = await download_files(download_list, task_uuid) return IO.NodeOutput(model_path, file_3d) class Rodin3D_Gen2(IO.ComfyNode): """Generate 3D Assets using Rodin API""" @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id="Rodin3D_Gen2", display_name="Rodin 3D Generate - Gen-2 Generate", category="api node/3d/Rodin", description=cleandoc(cls.__doc__ or ""), inputs=[ IO.Image.Input("Images"), IO.Int.Input( "Seed", default=0, min=0, max=65535, display_mode=IO.NumberDisplay.number, optional=True, ), IO.Combo.Input("Material_Type", options=["PBR", "Shaded"], default="PBR", optional=True), IO.Combo.Input( "Polygon_count", options=["4K-Quad", "8K-Quad", "18K-Quad", "50K-Quad", "2K-Triangle", "20K-Triangle", "150K-Triangle", "500K-Triangle"], default="500K-Triangle", optional=True, ), IO.Boolean.Input("TAPose", default=False, advanced=True), ], outputs=[ IO.String.Output(display_name="3D Model Path"), # for backward compatibility only IO.File3DGLB.Output(display_name="GLB"), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.4}""", ), ) @classmethod async def execute( cls, Images, Seed, Material_Type, Polygon_count, TAPose, ) -> IO.NodeOutput: tier = "Gen-2" num_images = Images.shape[0] m_images = [] for i in range(num_images): m_images.append(Images[i]) mesh_mode, quality_override = get_quality_mode(Polygon_count) task_uuid, subscription_key = await create_generate_task( cls, images=m_images, seed=Seed, material=Material_Type, quality_override=quality_override, tier=tier, mesh_mode=mesh_mode, ta_pose=TAPose, ) await poll_for_task_status(subscription_key, cls) download_list = await get_rodin_download_list(task_uuid, cls) model_path, file_3d = await download_files(download_list, task_uuid) return IO.NodeOutput(model_path, file_3d) class Rodin3DExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[IO.ComfyNode]]: return [ Rodin3D_Regular, Rodin3D_Detail, Rodin3D_Smooth, Rodin3D_Sketch, Rodin3D_Gen2, ] async def comfy_entrypoint() -> Rodin3DExtension: return Rodin3DExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api_nodes/nodes_rodin.py", "license": "GNU General Public License v3.0", "lines": 497, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_api_nodes/nodes_runway.py
"""Runway API Nodes API Docs: - https://docs.dev.runwayml.com/api/#tag/Task-management/paths/~1v1~1tasks~1%7Bid%7D/delete User Guides: - https://help.runwayml.com/hc/en-us/sections/30265301423635-Gen-3-Alpha - https://help.runwayml.com/hc/en-us/articles/37327109429011-Creating-with-Gen-4-Video - https://help.runwayml.com/hc/en-us/articles/33927968552339-Creating-with-Act-One-on-Gen-3-Alpha-and-Turbo - https://help.runwayml.com/hc/en-us/articles/34170748696595-Creating-with-Keyframes-on-Gen-3 """ from enum import Enum from typing_extensions import override from comfy_api.latest import IO, ComfyExtension, Input, InputImpl from comfy_api_nodes.apis.runway import ( RunwayImageToVideoRequest, RunwayImageToVideoResponse, RunwayTaskStatusResponse as TaskStatusResponse, RunwayModelEnum as Model, RunwayDurationEnum as Duration, RunwayAspectRatioEnum as AspectRatio, RunwayPromptImageObject, RunwayPromptImageDetailedObject, RunwayTextToImageRequest, RunwayTextToImageResponse, Model4, ReferenceImage, RunwayTextToImageAspectRatioEnum, ) from comfy_api_nodes.util import ( image_tensor_pair_to_batch, validate_string, validate_image_dimensions, validate_image_aspect_ratio, upload_images_to_comfyapi, download_url_to_video_output, download_url_to_image_tensor, ApiEndpoint, sync_op, poll_op, ) PATH_IMAGE_TO_VIDEO = "/proxy/runway/image_to_video" PATH_TEXT_TO_IMAGE = "/proxy/runway/text_to_image" PATH_GET_TASK_STATUS = "/proxy/runway/tasks" AVERAGE_DURATION_I2V_SECONDS = 64 AVERAGE_DURATION_FLF_SECONDS = 256 AVERAGE_DURATION_T2I_SECONDS = 41 class RunwayApiError(Exception): """Base exception for Runway API errors.""" pass class RunwayGen4TurboAspectRatio(str, Enum): """Aspect ratios supported for Image to Video API when using gen4_turbo model.""" field_1280_720 = "1280:720" field_720_1280 = "720:1280" field_1104_832 = "1104:832" field_832_1104 = "832:1104" field_960_960 = "960:960" field_1584_672 = "1584:672" class RunwayGen3aAspectRatio(str, Enum): """Aspect ratios supported for Image to Video API when using gen3a_turbo model.""" field_768_1280 = "768:1280" field_1280_768 = "1280:768" def get_video_url_from_task_status(response: TaskStatusResponse) -> str | None: """Returns the video URL from the task status response if it exists.""" if hasattr(response, "output") and len(response.output) > 0: return response.output[0] return None def extract_progress_from_task_status( response: TaskStatusResponse, ) -> float | None: if hasattr(response, "progress") and response.progress is not None: return response.progress * 100 return None def get_image_url_from_task_status(response: TaskStatusResponse) -> str | None: """Returns the image URL from the task status response if it exists.""" if hasattr(response, "output") and len(response.output) > 0: return response.output[0] return None async def get_response( cls: type[IO.ComfyNode], task_id: str, estimated_duration: int | None = None ) -> TaskStatusResponse: """Poll the task status until it is finished then get the response.""" return await poll_op( cls, ApiEndpoint(path=f"{PATH_GET_TASK_STATUS}/{task_id}"), response_model=TaskStatusResponse, status_extractor=lambda r: r.status.value, estimated_duration=estimated_duration, progress_extractor=extract_progress_from_task_status, ) async def generate_video( cls: type[IO.ComfyNode], request: RunwayImageToVideoRequest, estimated_duration: int | None = None, ) -> InputImpl.VideoFromFile: initial_response = await sync_op( cls, endpoint=ApiEndpoint(path=PATH_IMAGE_TO_VIDEO, method="POST"), response_model=RunwayImageToVideoResponse, data=request, ) final_response = await get_response(cls, initial_response.id, estimated_duration) if not final_response.output: raise RunwayApiError("Runway task succeeded but no video data found in response.") video_url = get_video_url_from_task_status(final_response) return await download_url_to_video_output(video_url) class RunwayImageToVideoNodeGen3a(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="RunwayImageToVideoNodeGen3a", display_name="Runway Image to Video (Gen3a Turbo)", category="api node/video/Runway", description="Generate a video from a single starting frame using Gen3a Turbo model. " "Before diving in, review these best practices to ensure that " "your input selections will set your generation up for success: " "https://help.runwayml.com/hc/en-us/articles/33927968552339-Creating-with-Act-One-on-Gen-3-Alpha-and-Turbo.", inputs=[ IO.String.Input( "prompt", multiline=True, default="", tooltip="Text prompt for the generation", ), IO.Image.Input( "start_frame", tooltip="Start frame to be used for the video", ), IO.Combo.Input( "duration", options=Duration, ), IO.Combo.Input( "ratio", options=RunwayGen3aAspectRatio, ), IO.Int.Input( "seed", default=0, min=0, max=4294967295, step=1, control_after_generate=True, display_mode=IO.NumberDisplay.number, tooltip="Random seed for generation", ), ], outputs=[ IO.Video.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["duration"]), expr="""{"type":"usd","usd": 0.0715 * widgets.duration}""", ), ) @classmethod async def execute( cls, prompt: str, start_frame: Input.Image, duration: str, ratio: str, seed: int, ) -> IO.NodeOutput: validate_string(prompt, min_length=1) validate_image_dimensions(start_frame, max_width=7999, max_height=7999) validate_image_aspect_ratio(start_frame, (1, 2), (2, 1)) download_urls = await upload_images_to_comfyapi( cls, start_frame, max_images=1, mime_type="image/png", ) return IO.NodeOutput( await generate_video( cls, RunwayImageToVideoRequest( promptText=prompt, seed=seed, model=Model("gen3a_turbo"), duration=Duration(duration), ratio=AspectRatio(ratio), promptImage=RunwayPromptImageObject( root=[RunwayPromptImageDetailedObject(uri=str(download_urls[0]), position="first")] ), ), ) ) class RunwayImageToVideoNodeGen4(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="RunwayImageToVideoNodeGen4", display_name="Runway Image to Video (Gen4 Turbo)", category="api node/video/Runway", description="Generate a video from a single starting frame using Gen4 Turbo model. " "Before diving in, review these best practices to ensure that " "your input selections will set your generation up for success: " "https://help.runwayml.com/hc/en-us/articles/37327109429011-Creating-with-Gen-4-Video.", inputs=[ IO.String.Input( "prompt", multiline=True, default="", tooltip="Text prompt for the generation", ), IO.Image.Input( "start_frame", tooltip="Start frame to be used for the video", ), IO.Combo.Input( "duration", options=Duration, ), IO.Combo.Input( "ratio", options=RunwayGen4TurboAspectRatio, ), IO.Int.Input( "seed", default=0, min=0, max=4294967295, step=1, control_after_generate=True, display_mode=IO.NumberDisplay.number, tooltip="Random seed for generation", ), ], outputs=[ IO.Video.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["duration"]), expr="""{"type":"usd","usd": 0.0715 * widgets.duration}""", ), ) @classmethod async def execute( cls, prompt: str, start_frame: Input.Image, duration: str, ratio: str, seed: int, ) -> IO.NodeOutput: validate_string(prompt, min_length=1) validate_image_dimensions(start_frame, max_width=7999, max_height=7999) validate_image_aspect_ratio(start_frame, (1, 2), (2, 1)) download_urls = await upload_images_to_comfyapi( cls, start_frame, max_images=1, mime_type="image/png", ) return IO.NodeOutput( await generate_video( cls, RunwayImageToVideoRequest( promptText=prompt, seed=seed, model=Model("gen4_turbo"), duration=Duration(duration), ratio=AspectRatio(ratio), promptImage=RunwayPromptImageObject( root=[RunwayPromptImageDetailedObject(uri=str(download_urls[0]), position="first")] ), ), estimated_duration=AVERAGE_DURATION_FLF_SECONDS, ) ) class RunwayFirstLastFrameNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="RunwayFirstLastFrameNode", display_name="Runway First-Last-Frame to Video", category="api node/video/Runway", description="Upload first and last keyframes, draft a prompt, and generate a video. " "More complex transitions, such as cases where the Last frame is completely different " "from the First frame, may benefit from the longer 10s duration. " "This would give the generation more time to smoothly transition between the two inputs. " "Before diving in, review these best practices to ensure that your input selections " "will set your generation up for success: " "https://help.runwayml.com/hc/en-us/articles/34170748696595-Creating-with-Keyframes-on-Gen-3.", inputs=[ IO.String.Input( "prompt", multiline=True, default="", tooltip="Text prompt for the generation", ), IO.Image.Input( "start_frame", tooltip="Start frame to be used for the video", ), IO.Image.Input( "end_frame", tooltip="End frame to be used for the video. Supported for gen3a_turbo only.", ), IO.Combo.Input( "duration", options=Duration, ), IO.Combo.Input( "ratio", options=RunwayGen3aAspectRatio, ), IO.Int.Input( "seed", default=0, min=0, max=4294967295, step=1, control_after_generate=True, display_mode=IO.NumberDisplay.number, tooltip="Random seed for generation", ), ], outputs=[ IO.Video.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["duration"]), expr="""{"type":"usd","usd": 0.0715 * widgets.duration}""", ), ) @classmethod async def execute( cls, prompt: str, start_frame: Input.Image, end_frame: Input.Image, duration: str, ratio: str, seed: int, ) -> IO.NodeOutput: validate_string(prompt, min_length=1) validate_image_dimensions(start_frame, max_width=7999, max_height=7999) validate_image_dimensions(end_frame, max_width=7999, max_height=7999) validate_image_aspect_ratio(start_frame, (1, 2), (2, 1)) validate_image_aspect_ratio(end_frame, (1, 2), (2, 1)) stacked_input_images = image_tensor_pair_to_batch(start_frame, end_frame) download_urls = await upload_images_to_comfyapi( cls, stacked_input_images, max_images=2, mime_type="image/png", ) if len(download_urls) != 2: raise RunwayApiError("Failed to upload one or more images to comfy api.") return IO.NodeOutput( await generate_video( cls, RunwayImageToVideoRequest( promptText=prompt, seed=seed, model=Model("gen3a_turbo"), duration=Duration(duration), ratio=AspectRatio(ratio), promptImage=RunwayPromptImageObject( root=[ RunwayPromptImageDetailedObject(uri=str(download_urls[0]), position="first"), RunwayPromptImageDetailedObject(uri=str(download_urls[1]), position="last"), ] ), ), estimated_duration=AVERAGE_DURATION_FLF_SECONDS, ) ) class RunwayTextToImageNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="RunwayTextToImageNode", display_name="Runway Text to Image", category="api node/image/Runway", description="Generate an image from a text prompt using Runway's Gen 4 model. " "You can also include reference image to guide the generation.", inputs=[ IO.String.Input( "prompt", multiline=True, default="", tooltip="Text prompt for the generation", ), IO.Combo.Input( "ratio", options=[model.value for model in RunwayTextToImageAspectRatioEnum], ), IO.Image.Input( "reference_image", tooltip="Optional reference image to guide the generation", optional=True, ), ], outputs=[ IO.Image.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.11}""", ), ) @classmethod async def execute( cls, prompt: str, ratio: str, reference_image: Input.Image | None = None, ) -> IO.NodeOutput: validate_string(prompt, min_length=1) # Prepare reference images if provided reference_images = None if reference_image is not None: validate_image_dimensions(reference_image, max_width=7999, max_height=7999) validate_image_aspect_ratio(reference_image, (1, 2), (2, 1)) download_urls = await upload_images_to_comfyapi( cls, reference_image, max_images=1, mime_type="image/png", ) reference_images = [ReferenceImage(uri=str(download_urls[0]))] initial_response = await sync_op( cls, endpoint=ApiEndpoint(path=PATH_TEXT_TO_IMAGE, method="POST"), response_model=RunwayTextToImageResponse, data=RunwayTextToImageRequest( promptText=prompt, model=Model4.gen4_image, ratio=ratio, referenceImages=reference_images, ), ) final_response = await get_response( cls, initial_response.id, estimated_duration=AVERAGE_DURATION_T2I_SECONDS, ) if not final_response.output: raise RunwayApiError("Runway task succeeded but no image data found in response.") return IO.NodeOutput(await download_url_to_image_tensor(get_image_url_from_task_status(final_response))) class RunwayExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[IO.ComfyNode]]: return [ RunwayFirstLastFrameNode, RunwayImageToVideoNodeGen3a, RunwayImageToVideoNodeGen4, RunwayTextToImageNode, ] async def comfy_entrypoint() -> RunwayExtension: return RunwayExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api_nodes/nodes_runway.py", "license": "GNU General Public License v3.0", "lines": 475, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_api_nodes/nodes_tripo.py
from typing_extensions import override from comfy_api.latest import IO, ComfyExtension, Input from comfy_api_nodes.apis.tripo import ( TripoAnimateRetargetRequest, TripoAnimateRigRequest, TripoConvertModelRequest, TripoFileEmptyReference, TripoFileReference, TripoImageToModelRequest, TripoModelVersion, TripoMultiviewToModelRequest, TripoOrientation, TripoRefineModelRequest, TripoStyle, TripoTaskResponse, TripoTaskStatus, TripoTaskType, TripoTextToModelRequest, TripoTextureModelRequest, TripoUrlReference, ) from comfy_api_nodes.util import ( ApiEndpoint, download_url_to_file_3d, poll_op, sync_op, upload_images_to_comfyapi, ) def get_model_url_from_response(response: TripoTaskResponse) -> str: if response.data is not None: for key in ["pbr_model", "model", "base_model"]: if getattr(response.data.output, key, None) is not None: return getattr(response.data.output, key) raise RuntimeError(f"Failed to get model url from response: {response}") async def poll_until_finished( node_cls: type[IO.ComfyNode], response: TripoTaskResponse, average_duration: int | None = None, ) -> IO.NodeOutput: """Polls the Tripo API endpoint until the task reaches a terminal state, then returns the response.""" if response.code != 0: raise RuntimeError(f"Failed to generate mesh: {response.error}") task_id = response.data.task_id response_poll = await poll_op( node_cls, poll_endpoint=ApiEndpoint(path=f"/proxy/tripo/v2/openapi/task/{task_id}"), response_model=TripoTaskResponse, completed_statuses=[TripoTaskStatus.SUCCESS], failed_statuses=[ TripoTaskStatus.FAILED, TripoTaskStatus.CANCELLED, TripoTaskStatus.UNKNOWN, TripoTaskStatus.BANNED, TripoTaskStatus.EXPIRED, ], status_extractor=lambda x: x.data.status, progress_extractor=lambda x: x.data.progress, estimated_duration=average_duration, ) if response_poll.data.status == TripoTaskStatus.SUCCESS: url = get_model_url_from_response(response_poll) file_glb = await download_url_to_file_3d(url, "glb", task_id=task_id) return IO.NodeOutput(f"{task_id}.glb", task_id, file_glb) raise RuntimeError(f"Failed to generate mesh: {response_poll}") class TripoTextToModelNode(IO.ComfyNode): """ Generates 3D models synchronously based on a text prompt using Tripo's API. """ @classmethod def define_schema(cls): return IO.Schema( node_id="TripoTextToModelNode", display_name="Tripo: Text to Model", category="api node/3d/Tripo", inputs=[ IO.String.Input("prompt", multiline=True), IO.String.Input("negative_prompt", multiline=True, optional=True), IO.Combo.Input( "model_version", options=TripoModelVersion, default=TripoModelVersion.v2_5_20250123, optional=True ), IO.Combo.Input("style", options=TripoStyle, default="None", optional=True), IO.Boolean.Input("texture", default=True, optional=True), IO.Boolean.Input("pbr", default=True, optional=True), IO.Int.Input("image_seed", default=42, optional=True, advanced=True), IO.Int.Input("model_seed", default=42, optional=True, advanced=True), IO.Int.Input("texture_seed", default=42, optional=True, advanced=True), IO.Combo.Input("texture_quality", default="standard", options=["standard", "detailed"], optional=True, advanced=True), IO.Int.Input("face_limit", default=-1, min=-1, max=2000000, optional=True, advanced=True), IO.Boolean.Input("quad", default=False, optional=True, advanced=True), IO.Combo.Input("geometry_quality", default="standard", options=["standard", "detailed"], optional=True, advanced=True), ], outputs=[ IO.String.Output(display_name="model_file"), # for backward compatibility only IO.Custom("MODEL_TASK_ID").Output(display_name="model task_id"), IO.File3DGLB.Output(display_name="GLB"), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, is_output_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends( widgets=[ "model_version", "style", "texture", "pbr", "quad", "texture_quality", "geometry_quality", ], ), expr=""" ( $isV14 := $contains(widgets.model_version,"v1.4"); $style := widgets.style; $hasStyle := ($style != "" and $style != "none"); $withTexture := widgets.texture or widgets.pbr; $isHdTexture := (widgets.texture_quality = "detailed"); $isDetailedGeometry := (widgets.geometry_quality = "detailed"); $baseCredits := $isV14 ? 20 : ($withTexture ? 20 : 10); $credits := $baseCredits + ($hasStyle ? 5 : 0) + (widgets.quad ? 5 : 0) + ($isHdTexture ? 10 : 0) + ($isDetailedGeometry ? 20 : 0); {"type":"usd","usd": $round($credits * 0.01, 2)} ) """, ), ) @classmethod async def execute( cls, prompt: str, negative_prompt: str | None = None, model_version=None, style: str | None = None, texture: bool | None = None, pbr: bool | None = None, image_seed: int | None = None, model_seed: int | None = None, texture_seed: int | None = None, texture_quality: str | None = None, geometry_quality: str | None = None, face_limit: int | None = None, quad: bool | None = None, ) -> IO.NodeOutput: style_enum = None if style == "None" else style if not prompt: raise RuntimeError("Prompt is required") response = await sync_op( cls, endpoint=ApiEndpoint(path="/proxy/tripo/v2/openapi/task", method="POST"), response_model=TripoTaskResponse, data=TripoTextToModelRequest( type=TripoTaskType.TEXT_TO_MODEL, prompt=prompt, negative_prompt=negative_prompt if negative_prompt else None, model_version=model_version, style=style_enum, texture=texture, pbr=pbr, image_seed=image_seed, model_seed=model_seed, texture_seed=texture_seed, texture_quality=texture_quality, face_limit=face_limit if face_limit != -1 else None, geometry_quality=geometry_quality, auto_size=True, quad=quad, ), ) return await poll_until_finished(cls, response, average_duration=80) class TripoImageToModelNode(IO.ComfyNode): """ Generates 3D models synchronously based on a single image using Tripo's API. """ @classmethod def define_schema(cls): return IO.Schema( node_id="TripoImageToModelNode", display_name="Tripo: Image to Model", category="api node/3d/Tripo", inputs=[ IO.Image.Input("image"), IO.Combo.Input( "model_version", options=TripoModelVersion, tooltip="The model version to use for generation", optional=True, ), IO.Combo.Input("style", options=TripoStyle, default="None", optional=True), IO.Boolean.Input("texture", default=True, optional=True), IO.Boolean.Input("pbr", default=True, optional=True), IO.Int.Input("model_seed", default=42, optional=True, advanced=True), IO.Combo.Input( "orientation", options=TripoOrientation, default=TripoOrientation.DEFAULT, optional=True, advanced=True ), IO.Int.Input("texture_seed", default=42, optional=True, advanced=True), IO.Combo.Input("texture_quality", default="standard", options=["standard", "detailed"], optional=True, advanced=True), IO.Combo.Input( "texture_alignment", default="original_image", options=["original_image", "geometry"], optional=True, advanced=True ), IO.Int.Input("face_limit", default=-1, min=-1, max=500000, optional=True, advanced=True), IO.Boolean.Input("quad", default=False, optional=True, advanced=True), IO.Combo.Input("geometry_quality", default="standard", options=["standard", "detailed"], optional=True, advanced=True), ], outputs=[ IO.String.Output(display_name="model_file"), # for backward compatibility only IO.Custom("MODEL_TASK_ID").Output(display_name="model task_id"), IO.File3DGLB.Output(display_name="GLB"), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, is_output_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends( widgets=[ "model_version", "style", "texture", "pbr", "quad", "texture_quality", "geometry_quality", ], ), expr=""" ( $isV14 := $contains(widgets.model_version,"v1.4"); $style := widgets.style; $hasStyle := ($style != "" and $style != "none"); $withTexture := widgets.texture or widgets.pbr; $isHdTexture := (widgets.texture_quality = "detailed"); $isDetailedGeometry := (widgets.geometry_quality = "detailed"); $baseCredits := $isV14 ? 30 : ($withTexture ? 30 : 20); $credits := $baseCredits + ($hasStyle ? 5 : 0) + (widgets.quad ? 5 : 0) + ($isHdTexture ? 10 : 0) + ($isDetailedGeometry ? 20 : 0); {"type":"usd","usd": $round($credits * 0.01, 2)} ) """, ), ) @classmethod async def execute( cls, image: Input.Image, model_version: str | None = None, style: str | None = None, texture: bool | None = None, pbr: bool | None = None, model_seed: int | None = None, orientation=None, texture_seed: int | None = None, texture_quality: str | None = None, geometry_quality: str | None = None, texture_alignment: str | None = None, face_limit: int | None = None, quad: bool | None = None, ) -> IO.NodeOutput: style_enum = None if style == "None" else style if image is None: raise RuntimeError("Image is required") tripo_file = TripoFileReference( root=TripoUrlReference( url=(await upload_images_to_comfyapi(cls, image, max_images=1))[0], type="jpeg", ) ) response = await sync_op( cls, endpoint=ApiEndpoint(path="/proxy/tripo/v2/openapi/task", method="POST"), response_model=TripoTaskResponse, data=TripoImageToModelRequest( type=TripoTaskType.IMAGE_TO_MODEL, file=tripo_file, model_version=model_version, style=style_enum, texture=texture, pbr=pbr, model_seed=model_seed, orientation=orientation, geometry_quality=geometry_quality, texture_alignment=texture_alignment, texture_seed=texture_seed, texture_quality=texture_quality, face_limit=face_limit if face_limit != -1 else None, auto_size=True, quad=quad, ), ) return await poll_until_finished(cls, response, average_duration=80) class TripoMultiviewToModelNode(IO.ComfyNode): """ Generates 3D models synchronously based on up to four images (front, left, back, right) using Tripo's API. """ @classmethod def define_schema(cls): return IO.Schema( node_id="TripoMultiviewToModelNode", display_name="Tripo: Multiview to Model", category="api node/3d/Tripo", inputs=[ IO.Image.Input("image"), IO.Image.Input("image_left", optional=True), IO.Image.Input("image_back", optional=True), IO.Image.Input("image_right", optional=True), IO.Combo.Input( "model_version", options=TripoModelVersion, optional=True, tooltip="The model version to use for generation", ), IO.Combo.Input( "orientation", options=TripoOrientation, default=TripoOrientation.DEFAULT, optional=True, advanced=True, ), IO.Boolean.Input("texture", default=True, optional=True), IO.Boolean.Input("pbr", default=True, optional=True), IO.Int.Input("model_seed", default=42, optional=True, advanced=True), IO.Int.Input("texture_seed", default=42, optional=True, advanced=True), IO.Combo.Input("texture_quality", default="standard", options=["standard", "detailed"], optional=True, advanced=True), IO.Combo.Input( "texture_alignment", default="original_image", options=["original_image", "geometry"], optional=True, advanced=True ), IO.Int.Input("face_limit", default=-1, min=-1, max=500000, optional=True, advanced=True), IO.Boolean.Input("quad", default=False, optional=True, advanced=True), IO.Combo.Input("geometry_quality", default="standard", options=["standard", "detailed"], optional=True, advanced=True), ], outputs=[ IO.String.Output(display_name="model_file"), # for backward compatibility only IO.Custom("MODEL_TASK_ID").Output(display_name="model task_id"), IO.File3DGLB.Output(display_name="GLB"), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, is_output_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends( widgets=[ "model_version", "texture", "pbr", "quad", "texture_quality", "geometry_quality", ], ), expr=""" ( $isV14 := $contains(widgets.model_version,"v1.4"); $withTexture := widgets.texture or widgets.pbr; $isHdTexture := (widgets.texture_quality = "detailed"); $isDetailedGeometry := (widgets.geometry_quality = "detailed"); $baseCredits := $isV14 ? 30 : ($withTexture ? 30 : 20); $credits := $baseCredits + (widgets.quad ? 5 : 0) + ($isHdTexture ? 10 : 0) + ($isDetailedGeometry ? 20 : 0); {"type":"usd","usd": $round($credits * 0.01, 2)} ) """, ), ) @classmethod async def execute( cls, image: Input.Image, image_left: Input.Image | None = None, image_back: Input.Image | None = None, image_right: Input.Image | None = None, model_version: str | None = None, orientation: str | None = None, texture: bool | None = None, pbr: bool | None = None, model_seed: int | None = None, texture_seed: int | None = None, texture_quality: str | None = None, geometry_quality: str | None = None, texture_alignment: str | None = None, face_limit: int | None = None, quad: bool | None = None, ) -> IO.NodeOutput: if image is None: raise RuntimeError("front image for multiview is required") images = [] image_dict = {"image": image, "image_left": image_left, "image_back": image_back, "image_right": image_right} if image_left is None and image_back is None and image_right is None: raise RuntimeError("At least one of left, back, or right image must be provided for multiview") for image_name in ["image", "image_left", "image_back", "image_right"]: image_ = image_dict[image_name] if image_ is not None: images.append( TripoFileReference( root=TripoUrlReference( url=(await upload_images_to_comfyapi(cls, image_, max_images=1))[0], type="jpeg" ) ) ) else: images.append(TripoFileEmptyReference()) response = await sync_op( cls, ApiEndpoint(path="/proxy/tripo/v2/openapi/task", method="POST"), response_model=TripoTaskResponse, data=TripoMultiviewToModelRequest( type=TripoTaskType.MULTIVIEW_TO_MODEL, files=images, model_version=model_version, orientation=orientation, texture=texture, pbr=pbr, model_seed=model_seed, texture_seed=texture_seed, texture_quality=texture_quality, geometry_quality=geometry_quality, texture_alignment=texture_alignment, face_limit=face_limit if face_limit != -1 else None, quad=quad, ), ) return await poll_until_finished(cls, response, average_duration=80) class TripoTextureNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="TripoTextureNode", display_name="Tripo: Texture model", category="api node/3d/Tripo", inputs=[ IO.Custom("MODEL_TASK_ID").Input("model_task_id"), IO.Boolean.Input("texture", default=True, optional=True), IO.Boolean.Input("pbr", default=True, optional=True), IO.Int.Input("texture_seed", default=42, optional=True, advanced=True), IO.Combo.Input("texture_quality", default="standard", options=["standard", "detailed"], optional=True, advanced=True), IO.Combo.Input( "texture_alignment", default="original_image", options=["original_image", "geometry"], optional=True, advanced=True ), ], outputs=[ IO.String.Output(display_name="model_file"), # for backward compatibility only IO.Custom("MODEL_TASK_ID").Output(display_name="model task_id"), IO.File3DGLB.Output(display_name="GLB"), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, is_output_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["texture_quality"]), expr=""" ( $tq := widgets.texture_quality; {"type":"usd","usd": ($contains($tq,"detailed") ? 0.2 : 0.1)} ) """, ), ) @classmethod async def execute( cls, model_task_id, texture: bool | None = None, pbr: bool | None = None, texture_seed: int | None = None, texture_quality: str | None = None, texture_alignment: str | None = None, ) -> IO.NodeOutput: response = await sync_op( cls, endpoint=ApiEndpoint(path="/proxy/tripo/v2/openapi/task", method="POST"), response_model=TripoTaskResponse, data=TripoTextureModelRequest( original_model_task_id=model_task_id, texture=texture, pbr=pbr, texture_seed=texture_seed, texture_quality=texture_quality, texture_alignment=texture_alignment, ), ) return await poll_until_finished(cls, response, average_duration=80) class TripoRefineNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="TripoRefineNode", display_name="Tripo: Refine Draft model", category="api node/3d/Tripo", description="Refine a draft model created by v1.4 Tripo models only.", inputs=[ IO.Custom("MODEL_TASK_ID").Input("model_task_id", tooltip="Must be a v1.4 Tripo model"), ], outputs=[ IO.String.Output(display_name="model_file"), # for backward compatibility only IO.Custom("MODEL_TASK_ID").Output(display_name="model task_id"), IO.File3DGLB.Output(display_name="GLB"), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, is_output_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.3}""", ), ) @classmethod async def execute(cls, model_task_id) -> IO.NodeOutput: response = await sync_op( cls, endpoint=ApiEndpoint(path="/proxy/tripo/v2/openapi/task", method="POST"), response_model=TripoTaskResponse, data=TripoRefineModelRequest(draft_model_task_id=model_task_id), ) return await poll_until_finished(cls, response, average_duration=240) class TripoRigNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="TripoRigNode", display_name="Tripo: Rig model", category="api node/3d/Tripo", inputs=[IO.Custom("MODEL_TASK_ID").Input("original_model_task_id")], outputs=[ IO.String.Output(display_name="model_file"), # for backward compatibility only IO.Custom("RIG_TASK_ID").Output(display_name="rig task_id"), IO.File3DGLB.Output(display_name="GLB"), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, is_output_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.25}""", ), ) @classmethod async def execute(cls, original_model_task_id) -> IO.NodeOutput: response = await sync_op( cls, endpoint=ApiEndpoint(path="/proxy/tripo/v2/openapi/task", method="POST"), response_model=TripoTaskResponse, data=TripoAnimateRigRequest(original_model_task_id=original_model_task_id, out_format="glb", spec="tripo"), ) return await poll_until_finished(cls, response, average_duration=180) class TripoRetargetNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="TripoRetargetNode", display_name="Tripo: Retarget rigged model", category="api node/3d/Tripo", inputs=[ IO.Custom("RIG_TASK_ID").Input("original_model_task_id"), IO.Combo.Input( "animation", options=[ "preset:idle", "preset:walk", "preset:run", "preset:dive", "preset:climb", "preset:jump", "preset:slash", "preset:shoot", "preset:hurt", "preset:fall", "preset:turn", "preset:quadruped:walk", "preset:hexapod:walk", "preset:octopod:walk", "preset:serpentine:march", "preset:aquatic:march" ], ), ], outputs=[ IO.String.Output(display_name="model_file"), # for backward compatibility only IO.Custom("RETARGET_TASK_ID").Output(display_name="retarget task_id"), IO.File3DGLB.Output(display_name="GLB"), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, is_output_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.1}""", ), ) @classmethod async def execute(cls, original_model_task_id, animation: str) -> IO.NodeOutput: response = await sync_op( cls, endpoint=ApiEndpoint(path="/proxy/tripo/v2/openapi/task", method="POST"), response_model=TripoTaskResponse, data=TripoAnimateRetargetRequest( original_model_task_id=original_model_task_id, animation=animation, out_format="glb", bake_animation=True, ), ) return await poll_until_finished(cls, response, average_duration=30) class TripoConversionNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="TripoConversionNode", display_name="Tripo: Convert model", category="api node/3d/Tripo", inputs=[ IO.Custom("MODEL_TASK_ID,RIG_TASK_ID,RETARGET_TASK_ID").Input("original_model_task_id"), IO.Combo.Input("format", options=["GLTF", "USDZ", "FBX", "OBJ", "STL", "3MF"]), IO.Boolean.Input("quad", default=False, optional=True, advanced=True), IO.Int.Input( "face_limit", default=-1, min=-1, max=2000000, optional=True, advanced=True, ), IO.Int.Input( "texture_size", default=4096, min=128, max=4096, optional=True, advanced=True, ), IO.Combo.Input( "texture_format", options=["BMP", "DPX", "HDR", "JPEG", "OPEN_EXR", "PNG", "TARGA", "TIFF", "WEBP"], default="JPEG", optional=True, advanced=True, ), IO.Boolean.Input("force_symmetry", default=False, optional=True, advanced=True), IO.Boolean.Input("flatten_bottom", default=False, optional=True, advanced=True), IO.Float.Input( "flatten_bottom_threshold", default=0.0, min=0.0, max=1.0, optional=True, advanced=True, ), IO.Boolean.Input("pivot_to_center_bottom", default=False, optional=True, advanced=True), IO.Float.Input( "scale_factor", default=1.0, min=0.0, optional=True, advanced=True, ), IO.Boolean.Input("with_animation", default=False, optional=True, advanced=True), IO.Boolean.Input("pack_uv", default=False, optional=True, advanced=True), IO.Boolean.Input("bake", default=False, optional=True, advanced=True), IO.String.Input("part_names", default="", optional=True, advanced=True), # comma-separated list IO.Combo.Input( "fbx_preset", options=["blender", "mixamo", "3dsmax"], default="blender", optional=True, advanced=True, ), IO.Boolean.Input("export_vertex_colors", default=False, optional=True, advanced=True), IO.Combo.Input( "export_orientation", options=["align_image", "default"], default="default", optional=True, advanced=True, ), IO.Boolean.Input("animate_in_place", default=False, optional=True, advanced=True), ], outputs=[], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, is_output_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends( widgets=[ "quad", "face_limit", "texture_size", "texture_format", "force_symmetry", "flatten_bottom", "flatten_bottom_threshold", "pivot_to_center_bottom", "scale_factor", "with_animation", "pack_uv", "bake", "part_names", "fbx_preset", "export_vertex_colors", "export_orientation", "animate_in_place", ], ), expr=""" ( $face := (widgets.face_limit != null) ? widgets.face_limit : -1; $texSize := (widgets.texture_size != null) ? widgets.texture_size : 4096; $flatThresh := (widgets.flatten_bottom_threshold != null) ? widgets.flatten_bottom_threshold : 0; $scale := (widgets.scale_factor != null) ? widgets.scale_factor : 1; $texFmt := (widgets.texture_format != "" ? widgets.texture_format : "jpeg"); $part := widgets.part_names; $fbx := (widgets.fbx_preset != "" ? widgets.fbx_preset : "blender"); $orient := (widgets.export_orientation != "" ? widgets.export_orientation : "default"); $advanced := widgets.quad or widgets.force_symmetry or widgets.flatten_bottom or widgets.pivot_to_center_bottom or widgets.with_animation or widgets.pack_uv or widgets.bake or widgets.export_vertex_colors or widgets.animate_in_place or ($face != -1) or ($texSize != 4096) or ($flatThresh != 0) or ($scale != 1) or ($texFmt != "jpeg") or ($part != "") or ($fbx != "blender") or ($orient != "default"); {"type":"usd","usd": ($advanced ? 0.1 : 0.05)} ) """, ), ) @classmethod def validate_inputs(cls, input_types): # The min and max of input1 and input2 are still validated because # we didn't take `input1` or `input2` as arguments if input_types["original_model_task_id"] not in ("MODEL_TASK_ID", "RIG_TASK_ID", "RETARGET_TASK_ID"): return "original_model_task_id must be MODEL_TASK_ID, RIG_TASK_ID or RETARGET_TASK_ID type" return True @classmethod async def execute( cls, original_model_task_id, format: str, quad: bool, force_symmetry: bool, face_limit: int, flatten_bottom: bool, flatten_bottom_threshold: float, texture_size: int, texture_format: str, pivot_to_center_bottom: bool, scale_factor: float, with_animation: bool, pack_uv: bool, bake: bool, part_names: str, fbx_preset: str, export_vertex_colors: bool, export_orientation: str, animate_in_place: bool, ) -> IO.NodeOutput: if not original_model_task_id: raise RuntimeError("original_model_task_id is required") # Parse part_names from comma-separated string to list part_names_list = None if part_names and part_names.strip(): part_names_list = [name.strip() for name in part_names.split(',') if name.strip()] response = await sync_op( cls, endpoint=ApiEndpoint(path="/proxy/tripo/v2/openapi/task", method="POST"), response_model=TripoTaskResponse, data=TripoConvertModelRequest( original_model_task_id=original_model_task_id, format=format, quad=quad if quad else None, force_symmetry=force_symmetry if force_symmetry else None, face_limit=face_limit if face_limit != -1 else None, flatten_bottom=flatten_bottom if flatten_bottom else None, flatten_bottom_threshold=flatten_bottom_threshold if flatten_bottom_threshold != 0.0 else None, texture_size=texture_size if texture_size != 4096 else None, texture_format=texture_format if texture_format != "JPEG" else None, pivot_to_center_bottom=pivot_to_center_bottom if pivot_to_center_bottom else None, scale_factor=scale_factor if scale_factor != 1.0 else None, with_animation=with_animation if with_animation else None, pack_uv=pack_uv if pack_uv else None, bake=bake if bake else None, part_names=part_names_list, fbx_preset=fbx_preset if fbx_preset != "blender" else None, export_vertex_colors=export_vertex_colors if export_vertex_colors else None, export_orientation=export_orientation if export_orientation != "default" else None, animate_in_place=animate_in_place if animate_in_place else None, ), ) return await poll_until_finished(cls, response, average_duration=30) class TripoExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[IO.ComfyNode]]: return [ TripoTextToModelNode, TripoImageToModelNode, TripoMultiviewToModelNode, TripoTextureNode, TripoRefineNode, TripoRigNode, TripoRetargetNode, TripoConversionNode, ] async def comfy_entrypoint() -> TripoExtension: return TripoExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api_nodes/nodes_tripo.py", "license": "GNU General Public License v3.0", "lines": 853, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_api/torch_helpers/torch_compile.py
from __future__ import annotations import torch import comfy.utils from comfy.patcher_extension import WrappersMP from typing import TYPE_CHECKING, Callable, Optional if TYPE_CHECKING: from comfy.model_patcher import ModelPatcher from comfy.patcher_extension import WrapperExecutor COMPILE_KEY = "torch.compile" TORCH_COMPILE_KWARGS = "torch_compile_kwargs" def apply_torch_compile_factory(compiled_module_dict: dict[str, Callable]) -> Callable: ''' Create a wrapper that will refer to the compiled_diffusion_model. ''' def apply_torch_compile_wrapper(executor: WrapperExecutor, *args, **kwargs): try: orig_modules = {} for key, value in compiled_module_dict.items(): orig_modules[key] = comfy.utils.get_attr(executor.class_obj, key) comfy.utils.set_attr(executor.class_obj, key, value) return executor(*args, **kwargs) finally: for key, value in orig_modules.items(): comfy.utils.set_attr(executor.class_obj, key, value) return apply_torch_compile_wrapper def set_torch_compile_wrapper(model: ModelPatcher, backend: str, options: Optional[dict[str,str]]=None, mode: Optional[str]=None, fullgraph=False, dynamic: Optional[bool]=None, keys: list[str]=["diffusion_model"], *args, **kwargs): ''' Perform torch.compile that will be applied at sample time for either the whole model or specific params of the BaseModel instance. When keys is None, it will default to using ["diffusion_model"], compiling the whole diffusion_model. When a list of keys is provided, it will perform torch.compile on only the selected modules. ''' # clear out any other torch.compile wrappers model.remove_wrappers_with_key(WrappersMP.APPLY_MODEL, COMPILE_KEY) # if no keys, default to 'diffusion_model' if not keys: keys = ["diffusion_model"] # create kwargs dict that can be referenced later compile_kwargs = { "backend": backend, "options": options, "mode": mode, "fullgraph": fullgraph, "dynamic": dynamic, } # get a dict of compiled keys compiled_modules = {} for key in keys: compiled_modules[key] = torch.compile( model=model.get_model_object(key), **compile_kwargs, ) # add torch.compile wrapper wrapper_func = apply_torch_compile_factory( compiled_module_dict=compiled_modules, ) # store wrapper to run on BaseModel's apply_model function model.add_wrapper_with_key(WrappersMP.APPLY_MODEL, COMPILE_KEY, wrapper_func) # keep compile kwargs for reference model.model_options[TORCH_COMPILE_KWARGS] = compile_kwargs
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api/torch_helpers/torch_compile.py", "license": "GNU General Public License v3.0", "lines": 61, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy_api_nodes/util/validation_utils.py
import logging import torch from comfy_api.latest import Input def get_image_dimensions(image: torch.Tensor) -> tuple[int, int]: if len(image.shape) == 4: return image.shape[1], image.shape[2] elif len(image.shape) == 3: return image.shape[0], image.shape[1] else: raise ValueError("Invalid image tensor shape.") def validate_image_dimensions( image: torch.Tensor, min_width: int | None = None, max_width: int | None = None, min_height: int | None = None, max_height: int | None = None, ): height, width = get_image_dimensions(image) if min_width is not None and width < min_width: raise ValueError(f"Image width must be at least {min_width}px, got {width}px") if max_width is not None and width > max_width: raise ValueError(f"Image width must be at most {max_width}px, got {width}px") if min_height is not None and height < min_height: raise ValueError(f"Image height must be at least {min_height}px, got {height}px") if max_height is not None and height > max_height: raise ValueError(f"Image height must be at most {max_height}px, got {height}px") def validate_image_aspect_ratio( image: torch.Tensor, min_ratio: tuple[float, float] | None = None, # e.g. (1, 4) max_ratio: tuple[float, float] | None = None, # e.g. (4, 1) *, strict: bool = True, # True -> (min, max); False -> [min, max] ) -> float: """Validates that image aspect ratio is within min and max. If a bound is None, that side is not checked.""" w, h = get_image_dimensions(image) if w <= 0 or h <= 0: raise ValueError(f"Invalid image dimensions: {w}x{h}") ar = w / h _assert_ratio_bounds(ar, min_ratio=min_ratio, max_ratio=max_ratio, strict=strict) return ar def validate_images_aspect_ratio_closeness( first_image: torch.Tensor, second_image: torch.Tensor, min_rel: float, # e.g. 0.8 max_rel: float, # e.g. 1.25 *, strict: bool = False, # True -> (min, max); False -> [min, max] ) -> float: """ Validates that the two images' aspect ratios are 'close'. The closeness factor is C = max(ar1, ar2) / min(ar1, ar2) (C >= 1). We require C <= limit, where limit = max(max_rel, 1.0 / min_rel). Returns the computed closeness factor C. """ w1, h1 = get_image_dimensions(first_image) w2, h2 = get_image_dimensions(second_image) if min(w1, h1, w2, h2) <= 0: raise ValueError("Invalid image dimensions") ar1 = w1 / h1 ar2 = w2 / h2 closeness = max(ar1, ar2) / min(ar1, ar2) limit = max(max_rel, 1.0 / min_rel) if (closeness >= limit) if strict else (closeness > limit): raise ValueError( f"Aspect ratios must be close: ar1/ar2={ar1/ar2:.2g}, " f"allowed range {min_rel}–{max_rel} (limit {limit:.2g})." ) return closeness def validate_aspect_ratio_string( aspect_ratio: str, min_ratio: tuple[float, float] | None = None, # e.g. (1, 4) max_ratio: tuple[float, float] | None = None, # e.g. (4, 1) *, strict: bool = False, # True -> (min, max); False -> [min, max] ) -> float: """Parses 'X:Y' and validates it against optional bounds. Returns the numeric ratio.""" ar = _parse_aspect_ratio_string(aspect_ratio) _assert_ratio_bounds(ar, min_ratio=min_ratio, max_ratio=max_ratio, strict=strict) return ar def validate_video_dimensions( video: Input.Video, min_width: int | None = None, max_width: int | None = None, min_height: int | None = None, max_height: int | None = None, ): try: width, height = video.get_dimensions() except Exception as e: logging.error("Error getting dimensions of video: %s", e) return if min_width is not None and width < min_width: raise ValueError(f"Video width must be at least {min_width}px, got {width}px") if max_width is not None and width > max_width: raise ValueError(f"Video width must be at most {max_width}px, got {width}px") if min_height is not None and height < min_height: raise ValueError(f"Video height must be at least {min_height}px, got {height}px") if max_height is not None and height > max_height: raise ValueError(f"Video height must be at most {max_height}px, got {height}px") def validate_video_duration( video: Input.Video, min_duration: float | None = None, max_duration: float | None = None, ): try: duration = video.get_duration() except Exception as e: logging.error("Error getting duration of video: %s", e) return epsilon = 0.0001 if min_duration is not None and min_duration - epsilon > duration: raise ValueError(f"Video duration must be at least {min_duration}s, got {duration}s") if max_duration is not None and duration > max_duration + epsilon: raise ValueError(f"Video duration must be at most {max_duration}s, got {duration}s") def validate_video_frame_count( video: Input.Video, min_frame_count: int | None = None, max_frame_count: int | None = None, ): try: frame_count = video.get_frame_count() except Exception as e: logging.error("Error getting frame count of video: %s", e) return if min_frame_count is not None and min_frame_count > frame_count: raise ValueError(f"Video frame count must be at least {min_frame_count}, got {frame_count}") if max_frame_count is not None and frame_count > max_frame_count: raise ValueError(f"Video frame count must be at most {max_frame_count}, got {frame_count}") def get_number_of_images(images): if isinstance(images, torch.Tensor): return images.shape[0] if images.ndim >= 4 else 1 return len(images) def validate_audio_duration( audio: Input.Audio, min_duration: float | None = None, max_duration: float | None = None, ) -> None: sr = int(audio["sample_rate"]) dur = int(audio["waveform"].shape[-1]) / sr eps = 1.0 / sr if min_duration is not None and dur + eps < min_duration: raise ValueError(f"Audio duration must be at least {min_duration}s, got {dur + eps:.2f}s") if max_duration is not None and dur - eps > max_duration: raise ValueError(f"Audio duration must be at most {max_duration}s, got {dur - eps:.2f}s") def validate_string( string: str, strip_whitespace=True, field_name="prompt", min_length=None, max_length=None, ): if string is None: raise Exception(f"Field '{field_name}' cannot be empty.") if strip_whitespace: string = string.strip() if min_length and len(string) < min_length: raise Exception( f"Field '{field_name}' cannot be shorter than {min_length} characters; was {len(string)} characters long." ) if max_length and len(string) > max_length: raise Exception( f" Field '{field_name} cannot be longer than {max_length} characters; was {len(string)} characters long." ) def validate_container_format_is_mp4(video: Input.Video) -> None: """Validates video container format is MP4.""" container_format = video.get_container_format() if container_format not in ["mp4", "mov,mp4,m4a,3gp,3g2,mj2"]: raise ValueError(f"Only MP4 container format supported. Got: {container_format}") def _ratio_from_tuple(r: tuple[float, float]) -> float: a, b = r if a <= 0 or b <= 0: raise ValueError(f"Ratios must be positive, got {a}:{b}.") return a / b def _assert_ratio_bounds( ar: float, *, min_ratio: tuple[float, float] | None = None, max_ratio: tuple[float, float] | None = None, strict: bool = True, ) -> None: """Validate a numeric aspect ratio against optional min/max ratio bounds.""" lo = _ratio_from_tuple(min_ratio) if min_ratio is not None else None hi = _ratio_from_tuple(max_ratio) if max_ratio is not None else None if lo is not None and hi is not None and lo > hi: lo, hi = hi, lo # normalize order if caller swapped them if lo is not None: if (ar <= lo) if strict else (ar < lo): op = "<" if strict else "≤" raise ValueError(f"Aspect ratio `{ar:.2g}` must be {op} {lo:.2g}.") if hi is not None: if (ar >= hi) if strict else (ar > hi): op = "<" if strict else "≤" raise ValueError(f"Aspect ratio `{ar:.2g}` must be {op} {hi:.2g}.") def _parse_aspect_ratio_string(ar_str: str) -> float: """Parse 'X:Y' with integer parts into a positive float ratio X/Y.""" parts = ar_str.split(":") if len(parts) != 2: raise ValueError(f"Aspect ratio must be 'X:Y' (e.g., 16:9), got '{ar_str}'.") try: a = int(parts[0].strip()) b = int(parts[1].strip()) except ValueError as exc: raise ValueError(f"Aspect ratio must contain integers separated by ':', got '{ar_str}'.") from exc if a <= 0 or b <= 0: raise ValueError(f"Aspect ratio parts must be positive integers, got {a}:{b}.") return a / b
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api_nodes/util/validation_utils.py", "license": "GNU General Public License v3.0", "lines": 206, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_extras/nodes_camera_trajectory.py
import nodes import torch import numpy as np from einops import rearrange from typing_extensions import override import comfy.model_management from comfy_api.latest import ComfyExtension, io CAMERA_DICT = { "base_T_norm": 1.5, "base_angle": np.pi/3, "Static": { "angle":[0., 0., 0.], "T":[0., 0., 0.]}, "Pan Up": { "angle":[0., 0., 0.], "T":[0., -1., 0.]}, "Pan Down": { "angle":[0., 0., 0.], "T":[0.,1.,0.]}, "Pan Left": { "angle":[0., 0., 0.], "T":[-1.,0.,0.]}, "Pan Right": { "angle":[0., 0., 0.], "T": [1.,0.,0.]}, "Zoom In": { "angle":[0., 0., 0.], "T": [0.,0.,2.]}, "Zoom Out": { "angle":[0., 0., 0.], "T": [0.,0.,-2.]}, "Anti Clockwise (ACW)": { "angle": [0., 0., -1.], "T":[0., 0., 0.]}, "ClockWise (CW)": { "angle": [0., 0., 1.], "T":[0., 0., 0.]}, } def process_pose_params(cam_params, width=672, height=384, original_pose_width=1280, original_pose_height=720, device='cpu'): def get_relative_pose(cam_params): """Copied from https://github.com/hehao13/CameraCtrl/blob/main/inference.py """ abs_w2cs = [cam_param.w2c_mat for cam_param in cam_params] abs_c2ws = [cam_param.c2w_mat for cam_param in cam_params] cam_to_origin = 0 target_cam_c2w = np.array([ [1, 0, 0, 0], [0, 1, 0, -cam_to_origin], [0, 0, 1, 0], [0, 0, 0, 1] ]) abs2rel = target_cam_c2w @ abs_w2cs[0] ret_poses = [target_cam_c2w, ] + [abs2rel @ abs_c2w for abs_c2w in abs_c2ws[1:]] ret_poses = np.array(ret_poses, dtype=np.float32) return ret_poses """Modified from https://github.com/hehao13/CameraCtrl/blob/main/inference.py """ cam_params = [Camera(cam_param) for cam_param in cam_params] sample_wh_ratio = width / height pose_wh_ratio = original_pose_width / original_pose_height # Assuming placeholder ratios, change as needed if pose_wh_ratio > sample_wh_ratio: resized_ori_w = height * pose_wh_ratio for cam_param in cam_params: cam_param.fx = resized_ori_w * cam_param.fx / width else: resized_ori_h = width / pose_wh_ratio for cam_param in cam_params: cam_param.fy = resized_ori_h * cam_param.fy / height intrinsic = np.asarray([[cam_param.fx * width, cam_param.fy * height, cam_param.cx * width, cam_param.cy * height] for cam_param in cam_params], dtype=np.float32) K = torch.as_tensor(intrinsic)[None] # [1, 1, 4] c2ws = get_relative_pose(cam_params) # Assuming this function is defined elsewhere c2ws = torch.as_tensor(c2ws)[None] # [1, n_frame, 4, 4] plucker_embedding = ray_condition(K, c2ws, height, width, device=device)[0].permute(0, 3, 1, 2).contiguous() # V, 6, H, W plucker_embedding = plucker_embedding[None] plucker_embedding = rearrange(plucker_embedding, "b f c h w -> b f h w c")[0] return plucker_embedding class Camera(object): """Copied from https://github.com/hehao13/CameraCtrl/blob/main/inference.py """ def __init__(self, entry): fx, fy, cx, cy = entry[1:5] self.fx = fx self.fy = fy self.cx = cx self.cy = cy c2w_mat = np.array(entry[7:]).reshape(4, 4) self.c2w_mat = c2w_mat self.w2c_mat = np.linalg.inv(c2w_mat) def ray_condition(K, c2w, H, W, device): """Copied from https://github.com/hehao13/CameraCtrl/blob/main/inference.py """ # c2w: B, V, 4, 4 # K: B, V, 4 B = K.shape[0] j, i = torch.meshgrid( torch.linspace(0, H - 1, H, device=device, dtype=c2w.dtype), torch.linspace(0, W - 1, W, device=device, dtype=c2w.dtype), indexing='ij' ) i = i.reshape([1, 1, H * W]).expand([B, 1, H * W]) + 0.5 # [B, HxW] j = j.reshape([1, 1, H * W]).expand([B, 1, H * W]) + 0.5 # [B, HxW] fx, fy, cx, cy = K.chunk(4, dim=-1) # B,V, 1 zs = torch.ones_like(i) # [B, HxW] xs = (i - cx) / fx * zs ys = (j - cy) / fy * zs zs = zs.expand_as(ys) directions = torch.stack((xs, ys, zs), dim=-1) # B, V, HW, 3 directions = directions / directions.norm(dim=-1, keepdim=True) # B, V, HW, 3 rays_d = directions @ c2w[..., :3, :3].transpose(-1, -2) # B, V, 3, HW rays_o = c2w[..., :3, 3] # B, V, 3 rays_o = rays_o[:, :, None].expand_as(rays_d) # B, V, 3, HW # c2w @ dirctions rays_dxo = torch.cross(rays_o, rays_d) plucker = torch.cat([rays_dxo, rays_d], dim=-1) plucker = plucker.reshape(B, c2w.shape[1], H, W, 6) # B, V, H, W, 6 # plucker = plucker.permute(0, 1, 4, 2, 3) return plucker def get_camera_motion(angle, T, speed, n=81): def compute_R_form_rad_angle(angles): theta_x, theta_y, theta_z = angles Rx = np.array([[1, 0, 0], [0, np.cos(theta_x), -np.sin(theta_x)], [0, np.sin(theta_x), np.cos(theta_x)]]) Ry = np.array([[np.cos(theta_y), 0, np.sin(theta_y)], [0, 1, 0], [-np.sin(theta_y), 0, np.cos(theta_y)]]) Rz = np.array([[np.cos(theta_z), -np.sin(theta_z), 0], [np.sin(theta_z), np.cos(theta_z), 0], [0, 0, 1]]) R = np.dot(Rz, np.dot(Ry, Rx)) return R RT = [] for i in range(n): _angle = (i/n)*speed*(CAMERA_DICT["base_angle"])*angle R = compute_R_form_rad_angle(_angle) _T=(i/n)*speed*(CAMERA_DICT["base_T_norm"])*(T.reshape(3,1)) _RT = np.concatenate([R,_T], axis=1) RT.append(_RT) RT = np.stack(RT) return RT class WanCameraEmbedding(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="WanCameraEmbedding", category="camera", inputs=[ io.Combo.Input( "camera_pose", options=[ "Static", "Pan Up", "Pan Down", "Pan Left", "Pan Right", "Zoom In", "Zoom Out", "Anti Clockwise (ACW)", "ClockWise (CW)", ], default="Static", ), io.Int.Input("width", default=832, min=16, max=nodes.MAX_RESOLUTION, step=16), io.Int.Input("height", default=480, min=16, max=nodes.MAX_RESOLUTION, step=16), io.Int.Input("length", default=81, min=1, max=nodes.MAX_RESOLUTION, step=4), io.Float.Input("speed", default=1.0, min=0, max=10.0, step=0.1, optional=True), io.Float.Input("fx", default=0.5, min=0, max=1, step=0.000000001, optional=True, advanced=True), io.Float.Input("fy", default=0.5, min=0, max=1, step=0.000000001, optional=True, advanced=True), io.Float.Input("cx", default=0.5, min=0, max=1, step=0.01, optional=True, advanced=True), io.Float.Input("cy", default=0.5, min=0, max=1, step=0.01, optional=True, advanced=True), ], outputs=[ io.WanCameraEmbedding.Output(display_name="camera_embedding"), io.Int.Output(display_name="width"), io.Int.Output(display_name="height"), io.Int.Output(display_name="length"), ], ) @classmethod def execute(cls, camera_pose, width, height, length, speed=1.0, fx=0.5, fy=0.5, cx=0.5, cy=0.5) -> io.NodeOutput: """ Use Camera trajectory as extrinsic parameters to calculate Plücker embeddings (Sitzmannet al., 2021) Adapted from https://github.com/aigc-apps/VideoX-Fun/blob/main/comfyui/comfyui_nodes.py """ motion_list = [camera_pose] speed = speed angle = np.array(CAMERA_DICT[motion_list[0]]["angle"]) T = np.array(CAMERA_DICT[motion_list[0]]["T"]) RT = get_camera_motion(angle, T, speed, length) trajs=[] for cp in RT.tolist(): traj=[fx,fy,cx,cy,0,0] traj.extend(cp[0]) traj.extend(cp[1]) traj.extend(cp[2]) traj.extend([0,0,0,1]) trajs.append(traj) cam_params = np.array([[float(x) for x in pose] for pose in trajs]) cam_params = np.concatenate([np.zeros_like(cam_params[:, :1]), cam_params], 1) control_camera_video = process_pose_params(cam_params, width=width, height=height) control_camera_video = control_camera_video.permute([3, 0, 1, 2]).unsqueeze(0).to(device=comfy.model_management.intermediate_device()) control_camera_video = torch.concat( [ torch.repeat_interleave(control_camera_video[:, :, 0:1], repeats=4, dim=2), control_camera_video[:, :, 1:] ], dim=2 ).transpose(1, 2) # Reshape, transpose, and view into desired shape b, f, c, h, w = control_camera_video.shape control_camera_video = control_camera_video.contiguous().view(b, f // 4, 4, c, h, w).transpose(2, 3) control_camera_video = control_camera_video.contiguous().view(b, f // 4, c * 4, h, w).transpose(1, 2) return io.NodeOutput(control_camera_video, width, height, length) class CameraTrajectoryExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[io.ComfyNode]]: return [ WanCameraEmbedding, ] async def comfy_entrypoint() -> CameraTrajectoryExtension: return CameraTrajectoryExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_extras/nodes_camera_trajectory.py", "license": "GNU General Public License v3.0", "lines": 206, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:tests-unit/comfy_api_test/video_types_test.py
import pytest import torch import tempfile import os import av import io from fractions import Fraction from comfy_api.input_impl.video_types import VideoFromFile, VideoFromComponents from comfy_api.util.video_types import VideoComponents from comfy_api.input.basic_types import AudioInput from av.error import InvalidDataError EPSILON = 0.0001 @pytest.fixture def sample_images(): """3-frame 2x2 RGB video tensor""" return torch.rand(3, 2, 2, 3) @pytest.fixture def sample_audio(): """Stereo audio with 44.1kHz sample rate""" return AudioInput( { "waveform": torch.rand(1, 2, 1000), "sample_rate": 44100, } ) @pytest.fixture def video_components(sample_images, sample_audio): """VideoComponents with images, audio, and metadata""" return VideoComponents( images=sample_images, audio=sample_audio, frame_rate=Fraction(30), metadata={"test": "metadata"}, ) def create_test_video(width=4, height=4, frames=3, fps=30): """Helper to create a temporary video file""" tmp = tempfile.NamedTemporaryFile(suffix=".mp4", delete=False) with av.open(tmp.name, mode="w") as container: stream = container.add_stream("h264", rate=fps) stream.width = width stream.height = height stream.pix_fmt = "yuv420p" for i in range(frames): frame = av.VideoFrame.from_ndarray( torch.ones(height, width, 3, dtype=torch.uint8).numpy() * (i * 85), format="rgb24", ) frame = frame.reformat(format="yuv420p") packet = stream.encode(frame) container.mux(packet) # Flush packet = stream.encode(None) container.mux(packet) return tmp.name @pytest.fixture def simple_video_file(): """4x4 video with 3 frames at 30fps""" file_path = create_test_video() yield file_path os.unlink(file_path) def test_video_from_components_get_duration(video_components): """Duration calculated correctly from frame count and frame rate""" video = VideoFromComponents(video_components) duration = video.get_duration() expected_duration = 3.0 / 30.0 assert duration == pytest.approx(expected_duration) def test_video_from_components_get_duration_different_frame_rates(sample_images): """Duration correct for different frame rates including fractional""" # Test with 60 fps components_60fps = VideoComponents(images=sample_images, frame_rate=Fraction(60)) video_60fps = VideoFromComponents(components_60fps) assert video_60fps.get_duration() == pytest.approx(3.0 / 60.0) # Test with fractional frame rate (23.976fps) components_frac = VideoComponents( images=sample_images, frame_rate=Fraction(24000, 1001) ) video_frac = VideoFromComponents(components_frac) expected_frac = 3.0 / (24000.0 / 1001.0) assert video_frac.get_duration() == pytest.approx(expected_frac) def test_video_from_components_get_duration_empty_video(): """Duration is zero for empty video""" empty_components = VideoComponents( images=torch.zeros(0, 2, 2, 3), frame_rate=Fraction(30) ) video = VideoFromComponents(empty_components) assert video.get_duration() == 0.0 def test_video_from_components_get_dimensions(video_components): """Dimensions returned correctly from image tensor shape""" video = VideoFromComponents(video_components) width, height = video.get_dimensions() assert width == 2 assert height == 2 def test_video_from_file_get_duration(simple_video_file): """Duration extracted from file metadata""" video = VideoFromFile(simple_video_file) duration = video.get_duration() assert duration == pytest.approx(0.1, abs=0.01) def test_video_from_file_get_dimensions(simple_video_file): """Dimensions read from stream without decoding frames""" video = VideoFromFile(simple_video_file) width, height = video.get_dimensions() assert width == 4 assert height == 4 def test_video_from_file_bytesio_input(): """VideoFromFile works with BytesIO input""" buffer = io.BytesIO() with av.open(buffer, mode="w", format="mp4") as container: stream = container.add_stream("h264", rate=30) stream.width = 2 stream.height = 2 stream.pix_fmt = "yuv420p" frame = av.VideoFrame.from_ndarray( torch.zeros(2, 2, 3, dtype=torch.uint8).numpy(), format="rgb24" ) frame = frame.reformat(format="yuv420p") packet = stream.encode(frame) container.mux(packet) packet = stream.encode(None) container.mux(packet) buffer.seek(0) video = VideoFromFile(buffer) assert video.get_dimensions() == (2, 2) assert video.get_duration() == pytest.approx(1 / 30, abs=0.01) def test_video_from_file_invalid_file_error(): """InvalidDataError raised for non-video files""" with tempfile.NamedTemporaryFile(suffix=".txt", delete=False) as tmp: tmp.write(b"not a video file") tmp.flush() tmp_name = tmp.name try: with pytest.raises(InvalidDataError): video = VideoFromFile(tmp_name) video.get_dimensions() finally: os.unlink(tmp_name) def test_video_from_file_audio_only_error(): """ValueError raised for audio-only files""" with tempfile.NamedTemporaryFile(suffix=".m4a", delete=False) as tmp: tmp_name = tmp.name try: with av.open(tmp_name, mode="w") as container: stream = container.add_stream("aac", rate=44100) stream.sample_rate = 44100 stream.format = "fltp" audio_data = torch.zeros(1, 1024).numpy() audio_frame = av.AudioFrame.from_ndarray( audio_data, format="fltp", layout="mono" ) audio_frame.sample_rate = 44100 audio_frame.pts = 0 packet = stream.encode(audio_frame) container.mux(packet) for packet in stream.encode(None): container.mux(packet) with pytest.raises(ValueError, match="No video stream found"): video = VideoFromFile(tmp_name) video.get_dimensions() finally: os.unlink(tmp_name) def test_single_frame_video(): """Single frame video has correct duration""" components = VideoComponents( images=torch.rand(1, 10, 10, 3), frame_rate=Fraction(1) ) video = VideoFromComponents(components) assert video.get_duration() == 1.0 @pytest.mark.parametrize( "frame_rate,expected_fps", [ (Fraction(24000, 1001), 24000 / 1001), (Fraction(30000, 1001), 30000 / 1001), (Fraction(25, 1), 25.0), (Fraction(50, 2), 25.0), ], ) def test_fractional_frame_rates(frame_rate, expected_fps): """Duration calculated correctly for various fractional frame rates""" components = VideoComponents(images=torch.rand(100, 4, 4, 3), frame_rate=frame_rate) video = VideoFromComponents(components) duration = video.get_duration() expected_duration = 100.0 / expected_fps assert duration == pytest.approx(expected_duration) def test_duration_consistency(video_components): """get_duration() consistent with manual calculation from components""" video = VideoFromComponents(video_components) duration = video.get_duration() components = video.get_components() manual_duration = float(components.images.shape[0] / components.frame_rate) assert duration == pytest.approx(manual_duration)
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "tests-unit/comfy_api_test/video_types_test.py", "license": "GNU General Public License v3.0", "lines": 189, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
Comfy-Org/ComfyUI:comfy_extras/nodes_apg.py
import torch from typing_extensions import override from comfy_api.latest import ComfyExtension, io def project(v0, v1): v1 = torch.nn.functional.normalize(v1, dim=[-1, -2, -3]) v0_parallel = (v0 * v1).sum(dim=[-1, -2, -3], keepdim=True) * v1 v0_orthogonal = v0 - v0_parallel return v0_parallel, v0_orthogonal class APG(io.ComfyNode): @classmethod def define_schema(cls) -> io.Schema: return io.Schema( node_id="APG", display_name="Adaptive Projected Guidance", category="sampling/custom_sampling", inputs=[ io.Model.Input("model"), io.Float.Input( "eta", default=1.0, min=-10.0, max=10.0, step=0.01, tooltip="Controls the scale of the parallel guidance vector. Default CFG behavior at a setting of 1.", advanced=True, ), io.Float.Input( "norm_threshold", default=5.0, min=0.0, max=50.0, step=0.1, tooltip="Normalize guidance vector to this value, normalization disable at a setting of 0.", advanced=True, ), io.Float.Input( "momentum", default=0.0, min=-5.0, max=1.0, step=0.01, tooltip="Controls a running average of guidance during diffusion, disabled at a setting of 0.", advanced=True, ), ], outputs=[io.Model.Output()], ) @classmethod def execute(cls, model, eta, norm_threshold, momentum) -> io.NodeOutput: running_avg = 0 prev_sigma = None def pre_cfg_function(args): nonlocal running_avg, prev_sigma if len(args["conds_out"]) == 1: return args["conds_out"] cond = args["conds_out"][0] uncond = args["conds_out"][1] sigma = args["sigma"][0] cond_scale = args["cond_scale"] if prev_sigma is not None and sigma > prev_sigma: running_avg = 0 prev_sigma = sigma guidance = cond - uncond if momentum != 0: if not torch.is_tensor(running_avg): running_avg = guidance else: running_avg = momentum * running_avg + guidance guidance = running_avg if norm_threshold > 0: guidance_norm = guidance.norm(p=2, dim=[-1, -2, -3], keepdim=True) scale = torch.minimum( torch.ones_like(guidance_norm), norm_threshold / guidance_norm ) guidance = guidance * scale guidance_parallel, guidance_orthogonal = project(guidance, cond) modified_guidance = guidance_orthogonal + eta * guidance_parallel modified_cond = (uncond + modified_guidance) + (cond - uncond) / cond_scale return [modified_cond, uncond] + args["conds_out"][2:] m = model.clone() m.set_model_sampler_pre_cfg_function(pre_cfg_function) return io.NodeOutput(m) class ApgExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[io.ComfyNode]]: return [ APG, ] async def comfy_entrypoint() -> ApgExtension: return ApgExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_extras/nodes_apg.py", "license": "GNU General Public License v3.0", "lines": 91, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy_extras/nodes_string.py
import re from typing_extensions import override from comfy_api.latest import ComfyExtension, io class StringConcatenate(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="StringConcatenate", display_name="Concatenate", category="utils/string", search_aliases=["text concat", "join text", "merge text", "combine strings", "concat", "concatenate", "append text", "combine text", "string"], inputs=[ io.String.Input("string_a", multiline=True), io.String.Input("string_b", multiline=True), io.String.Input("delimiter", multiline=False, default=""), ], outputs=[ io.String.Output(), ] ) @classmethod def execute(cls, string_a, string_b, delimiter): return io.NodeOutput(delimiter.join((string_a, string_b))) class StringSubstring(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="StringSubstring", search_aliases=["extract text", "text portion"], display_name="Substring", category="utils/string", inputs=[ io.String.Input("string", multiline=True), io.Int.Input("start"), io.Int.Input("end"), ], outputs=[ io.String.Output(), ] ) @classmethod def execute(cls, string, start, end): return io.NodeOutput(string[start:end]) class StringLength(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="StringLength", search_aliases=["character count", "text size"], display_name="Length", category="utils/string", inputs=[ io.String.Input("string", multiline=True), ], outputs=[ io.Int.Output(display_name="length"), ] ) @classmethod def execute(cls, string): return io.NodeOutput(len(string)) class CaseConverter(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="CaseConverter", search_aliases=["text case", "uppercase", "lowercase", "capitalize"], display_name="Case Converter", category="utils/string", inputs=[ io.String.Input("string", multiline=True), io.Combo.Input("mode", options=["UPPERCASE", "lowercase", "Capitalize", "Title Case"]), ], outputs=[ io.String.Output(), ] ) @classmethod def execute(cls, string, mode): if mode == "UPPERCASE": result = string.upper() elif mode == "lowercase": result = string.lower() elif mode == "Capitalize": result = string.capitalize() elif mode == "Title Case": result = string.title() else: result = string return io.NodeOutput(result) class StringTrim(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="StringTrim", search_aliases=["clean whitespace", "remove whitespace"], display_name="Trim", category="utils/string", inputs=[ io.String.Input("string", multiline=True), io.Combo.Input("mode", options=["Both", "Left", "Right"]), ], outputs=[ io.String.Output(), ] ) @classmethod def execute(cls, string, mode): if mode == "Both": result = string.strip() elif mode == "Left": result = string.lstrip() elif mode == "Right": result = string.rstrip() else: result = string return io.NodeOutput(result) class StringReplace(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="StringReplace", search_aliases=["find and replace", "substitute", "swap text"], display_name="Replace", category="utils/string", inputs=[ io.String.Input("string", multiline=True), io.String.Input("find", multiline=True), io.String.Input("replace", multiline=True), ], outputs=[ io.String.Output(), ] ) @classmethod def execute(cls, string, find, replace): return io.NodeOutput(string.replace(find, replace)) class StringContains(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="StringContains", search_aliases=["text includes", "string includes"], display_name="Contains", category="utils/string", inputs=[ io.String.Input("string", multiline=True), io.String.Input("substring", multiline=True), io.Boolean.Input("case_sensitive", default=True, advanced=True), ], outputs=[ io.Boolean.Output(display_name="contains"), ] ) @classmethod def execute(cls, string, substring, case_sensitive): if case_sensitive: contains = substring in string else: contains = substring.lower() in string.lower() return io.NodeOutput(contains) class StringCompare(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="StringCompare", search_aliases=["text match", "string equals", "starts with", "ends with"], display_name="Compare", category="utils/string", inputs=[ io.String.Input("string_a", multiline=True), io.String.Input("string_b", multiline=True), io.Combo.Input("mode", options=["Starts With", "Ends With", "Equal"]), io.Boolean.Input("case_sensitive", default=True, advanced=True), ], outputs=[ io.Boolean.Output(), ] ) @classmethod def execute(cls, string_a, string_b, mode, case_sensitive): if case_sensitive: a = string_a b = string_b else: a = string_a.lower() b = string_b.lower() if mode == "Equal": return io.NodeOutput(a == b) elif mode == "Starts With": return io.NodeOutput(a.startswith(b)) elif mode == "Ends With": return io.NodeOutput(a.endswith(b)) class RegexMatch(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="RegexMatch", search_aliases=["pattern match", "text contains", "string match"], display_name="Regex Match", category="utils/string", inputs=[ io.String.Input("string", multiline=True), io.String.Input("regex_pattern", multiline=True), io.Boolean.Input("case_insensitive", default=True, advanced=True), io.Boolean.Input("multiline", default=False, advanced=True), io.Boolean.Input("dotall", default=False, advanced=True), ], outputs=[ io.Boolean.Output(display_name="matches"), ] ) @classmethod def execute(cls, string, regex_pattern, case_insensitive, multiline, dotall): flags = 0 if case_insensitive: flags |= re.IGNORECASE if multiline: flags |= re.MULTILINE if dotall: flags |= re.DOTALL try: match = re.search(regex_pattern, string, flags) result = match is not None except re.error: result = False return io.NodeOutput(result) class RegexExtract(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="RegexExtract", search_aliases=["pattern extract", "text parser", "parse text"], display_name="Regex Extract", category="utils/string", inputs=[ io.String.Input("string", multiline=True), io.String.Input("regex_pattern", multiline=True), io.Combo.Input("mode", options=["First Match", "All Matches", "First Group", "All Groups"]), io.Boolean.Input("case_insensitive", default=True, advanced=True), io.Boolean.Input("multiline", default=False, advanced=True), io.Boolean.Input("dotall", default=False, advanced=True), io.Int.Input("group_index", default=1, min=0, max=100, advanced=True), ], outputs=[ io.String.Output(), ] ) @classmethod def execute(cls, string, regex_pattern, mode, case_insensitive, multiline, dotall, group_index): join_delimiter = "\n" flags = 0 if case_insensitive: flags |= re.IGNORECASE if multiline: flags |= re.MULTILINE if dotall: flags |= re.DOTALL try: if mode == "First Match": match = re.search(regex_pattern, string, flags) if match: result = match.group(0) else: result = "" elif mode == "All Matches": matches = re.findall(regex_pattern, string, flags) if matches: if isinstance(matches[0], tuple): result = join_delimiter.join([m[0] for m in matches]) else: result = join_delimiter.join(matches) else: result = "" elif mode == "First Group": match = re.search(regex_pattern, string, flags) if match and len(match.groups()) >= group_index: result = match.group(group_index) else: result = "" elif mode == "All Groups": matches = re.finditer(regex_pattern, string, flags) results = [] for match in matches: if match.groups() and len(match.groups()) >= group_index: results.append(match.group(group_index)) result = join_delimiter.join(results) else: result = "" except re.error: result = "" return io.NodeOutput(result) class RegexReplace(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="RegexReplace", search_aliases=["pattern replace", "find and replace", "substitution"], display_name="Regex Replace", category="utils/string", description="Find and replace text using regex patterns.", inputs=[ io.String.Input("string", multiline=True), io.String.Input("regex_pattern", multiline=True), io.String.Input("replace", multiline=True), io.Boolean.Input("case_insensitive", default=True, optional=True, advanced=True), io.Boolean.Input("multiline", default=False, optional=True, advanced=True), io.Boolean.Input("dotall", default=False, optional=True, advanced=True, tooltip="When enabled, the dot (.) character will match any character including newline characters. When disabled, dots won't match newlines."), io.Int.Input("count", default=0, min=0, max=100, optional=True, advanced=True, tooltip="Maximum number of replacements to make. Set to 0 to replace all occurrences (default). Set to 1 to replace only the first match, 2 for the first two matches, etc."), ], outputs=[ io.String.Output(), ] ) @classmethod def execute(cls, string, regex_pattern, replace, case_insensitive=True, multiline=False, dotall=False, count=0): flags = 0 if case_insensitive: flags |= re.IGNORECASE if multiline: flags |= re.MULTILINE if dotall: flags |= re.DOTALL result = re.sub(regex_pattern, replace, string, count=count, flags=flags) return io.NodeOutput(result) class StringExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[io.ComfyNode]]: return [ StringConcatenate, StringSubstring, StringLength, CaseConverter, StringTrim, StringReplace, StringContains, StringCompare, RegexMatch, RegexExtract, RegexReplace, ] async def comfy_entrypoint() -> StringExtension: return StringExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_extras/nodes_string.py", "license": "GNU General Public License v3.0", "lines": 343, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy/ldm/ace/lyric_encoder.py
# Original from: https://github.com/ace-step/ACE-Step/blob/main/models/lyrics_utils/lyric_encoder.py from typing import Optional, Tuple, Union import math import torch from torch import nn import comfy.model_management class ConvolutionModule(nn.Module): """ConvolutionModule in Conformer model.""" def __init__(self, channels: int, kernel_size: int = 15, activation: nn.Module = nn.ReLU(), norm: str = "batch_norm", causal: bool = False, bias: bool = True, dtype=None, device=None, operations=None): """Construct an ConvolutionModule object. Args: channels (int): The number of channels of conv layers. kernel_size (int): Kernel size of conv layers. causal (int): Whether use causal convolution or not """ super().__init__() self.pointwise_conv1 = operations.Conv1d( channels, 2 * channels, kernel_size=1, stride=1, padding=0, bias=bias, dtype=dtype, device=device ) # self.lorder is used to distinguish if it's a causal convolution, # if self.lorder > 0: it's a causal convolution, the input will be # padded with self.lorder frames on the left in forward. # else: it's a symmetrical convolution if causal: padding = 0 self.lorder = kernel_size - 1 else: # kernel_size should be an odd number for none causal convolution assert (kernel_size - 1) % 2 == 0 padding = (kernel_size - 1) // 2 self.lorder = 0 self.depthwise_conv = operations.Conv1d( channels, channels, kernel_size, stride=1, padding=padding, groups=channels, bias=bias, dtype=dtype, device=device ) assert norm in ['batch_norm', 'layer_norm'] if norm == "batch_norm": self.use_layer_norm = False self.norm = nn.BatchNorm1d(channels) else: self.use_layer_norm = True self.norm = operations.LayerNorm(channels, dtype=dtype, device=device) self.pointwise_conv2 = operations.Conv1d( channels, channels, kernel_size=1, stride=1, padding=0, bias=bias, dtype=dtype, device=device ) self.activation = activation def forward( self, x: torch.Tensor, mask_pad: torch.Tensor = torch.ones((0, 0, 0), dtype=torch.bool), cache: torch.Tensor = torch.zeros((0, 0, 0)), ) -> Tuple[torch.Tensor, torch.Tensor]: """Compute convolution module. Args: x (torch.Tensor): Input tensor (#batch, time, channels). mask_pad (torch.Tensor): used for batch padding (#batch, 1, time), (0, 0, 0) means fake mask. cache (torch.Tensor): left context cache, it is only used in causal convolution (#batch, channels, cache_t), (0, 0, 0) meas fake cache. Returns: torch.Tensor: Output tensor (#batch, time, channels). """ # exchange the temporal dimension and the feature dimension x = x.transpose(1, 2) # (#batch, channels, time) # mask batch padding if mask_pad.size(2) > 0: # time > 0 x.masked_fill_(~mask_pad, 0.0) if self.lorder > 0: if cache.size(2) == 0: # cache_t == 0 x = nn.functional.pad(x, (self.lorder, 0), 'constant', 0.0) else: assert cache.size(0) == x.size(0) # equal batch assert cache.size(1) == x.size(1) # equal channel x = torch.cat((cache, x), dim=2) assert (x.size(2) > self.lorder) new_cache = x[:, :, -self.lorder:] else: # It's better we just return None if no cache is required, # However, for JIT export, here we just fake one tensor instead of # None. new_cache = torch.zeros((0, 0, 0), dtype=x.dtype, device=x.device) # GLU mechanism x = self.pointwise_conv1(x) # (batch, 2*channel, dim) x = nn.functional.glu(x, dim=1) # (batch, channel, dim) # 1D Depthwise Conv x = self.depthwise_conv(x) if self.use_layer_norm: x = x.transpose(1, 2) x = self.activation(self.norm(x)) if self.use_layer_norm: x = x.transpose(1, 2) x = self.pointwise_conv2(x) # mask batch padding if mask_pad.size(2) > 0: # time > 0 x.masked_fill_(~mask_pad, 0.0) return x.transpose(1, 2), new_cache class PositionwiseFeedForward(torch.nn.Module): """Positionwise feed forward layer. FeedForward are appied on each position of the sequence. The output dim is same with the input dim. Args: idim (int): Input dimenstion. hidden_units (int): The number of hidden units. dropout_rate (float): Dropout rate. activation (torch.nn.Module): Activation function """ def __init__( self, idim: int, hidden_units: int, dropout_rate: float, activation: torch.nn.Module = torch.nn.ReLU(), dtype=None, device=None, operations=None ): """Construct a PositionwiseFeedForward object.""" super(PositionwiseFeedForward, self).__init__() self.w_1 = operations.Linear(idim, hidden_units, dtype=dtype, device=device) self.activation = activation self.dropout = torch.nn.Dropout(dropout_rate) self.w_2 = operations.Linear(hidden_units, idim, dtype=dtype, device=device) def forward(self, xs: torch.Tensor) -> torch.Tensor: """Forward function. Args: xs: input tensor (B, L, D) Returns: output tensor, (B, L, D) """ return self.w_2(self.dropout(self.activation(self.w_1(xs)))) class Swish(torch.nn.Module): """Construct an Swish object.""" def forward(self, x: torch.Tensor) -> torch.Tensor: """Return Swish activation function.""" return x * torch.sigmoid(x) class MultiHeadedAttention(nn.Module): """Multi-Head Attention layer. Args: n_head (int): The number of heads. n_feat (int): The number of features. dropout_rate (float): Dropout rate. """ def __init__(self, n_head: int, n_feat: int, dropout_rate: float, key_bias: bool = True, dtype=None, device=None, operations=None): """Construct an MultiHeadedAttention object.""" super().__init__() assert n_feat % n_head == 0 # We assume d_v always equals d_k self.d_k = n_feat // n_head self.h = n_head self.linear_q = operations.Linear(n_feat, n_feat, dtype=dtype, device=device) self.linear_k = operations.Linear(n_feat, n_feat, bias=key_bias, dtype=dtype, device=device) self.linear_v = operations.Linear(n_feat, n_feat, dtype=dtype, device=device) self.linear_out = operations.Linear(n_feat, n_feat, dtype=dtype, device=device) self.dropout = nn.Dropout(p=dropout_rate) def forward_qkv( self, query: torch.Tensor, key: torch.Tensor, value: torch.Tensor ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: """Transform query, key and value. Args: query (torch.Tensor): Query tensor (#batch, time1, size). key (torch.Tensor): Key tensor (#batch, time2, size). value (torch.Tensor): Value tensor (#batch, time2, size). Returns: torch.Tensor: Transformed query tensor, size (#batch, n_head, time1, d_k). torch.Tensor: Transformed key tensor, size (#batch, n_head, time2, d_k). torch.Tensor: Transformed value tensor, size (#batch, n_head, time2, d_k). """ n_batch = query.size(0) q = self.linear_q(query).view(n_batch, -1, self.h, self.d_k) k = self.linear_k(key).view(n_batch, -1, self.h, self.d_k) v = self.linear_v(value).view(n_batch, -1, self.h, self.d_k) q = q.transpose(1, 2) # (batch, head, time1, d_k) k = k.transpose(1, 2) # (batch, head, time2, d_k) v = v.transpose(1, 2) # (batch, head, time2, d_k) return q, k, v def forward_attention( self, value: torch.Tensor, scores: torch.Tensor, mask: torch.Tensor = torch.ones((0, 0, 0), dtype=torch.bool) ) -> torch.Tensor: """Compute attention context vector. Args: value (torch.Tensor): Transformed value, size (#batch, n_head, time2, d_k). scores (torch.Tensor): Attention score, size (#batch, n_head, time1, time2). mask (torch.Tensor): Mask, size (#batch, 1, time2) or (#batch, time1, time2), (0, 0, 0) means fake mask. Returns: torch.Tensor: Transformed value (#batch, time1, d_model) weighted by the attention score (#batch, time1, time2). """ n_batch = value.size(0) if mask is not None and mask.size(2) > 0: # time2 > 0 mask = mask.unsqueeze(1).eq(0) # (batch, 1, *, time2) # For last chunk, time2 might be larger than scores.size(-1) mask = mask[:, :, :, :scores.size(-1)] # (batch, 1, *, time2) scores = scores.masked_fill(mask, -float('inf')) attn = torch.softmax(scores, dim=-1).masked_fill( mask, 0.0) # (batch, head, time1, time2) else: attn = torch.softmax(scores, dim=-1) # (batch, head, time1, time2) p_attn = self.dropout(attn) x = torch.matmul(p_attn, value) # (batch, head, time1, d_k) x = (x.transpose(1, 2).contiguous().view(n_batch, -1, self.h * self.d_k) ) # (batch, time1, d_model) return self.linear_out(x) # (batch, time1, d_model) def forward( self, query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, mask: torch.Tensor = torch.ones((0, 0, 0), dtype=torch.bool), pos_emb: torch.Tensor = torch.empty(0), cache: torch.Tensor = torch.zeros((0, 0, 0, 0)) ) -> Tuple[torch.Tensor, torch.Tensor]: """Compute scaled dot product attention. Args: query (torch.Tensor): Query tensor (#batch, time1, size). key (torch.Tensor): Key tensor (#batch, time2, size). value (torch.Tensor): Value tensor (#batch, time2, size). mask (torch.Tensor): Mask tensor (#batch, 1, time2) or (#batch, time1, time2). 1.When applying cross attention between decoder and encoder, the batch padding mask for input is in (#batch, 1, T) shape. 2.When applying self attention of encoder, the mask is in (#batch, T, T) shape. 3.When applying self attention of decoder, the mask is in (#batch, L, L) shape. 4.If the different position in decoder see different block of the encoder, such as Mocha, the passed in mask could be in (#batch, L, T) shape. But there is no such case in current CosyVoice. cache (torch.Tensor): Cache tensor (1, head, cache_t, d_k * 2), where `cache_t == chunk_size * num_decoding_left_chunks` and `head * d_k == size` Returns: torch.Tensor: Output tensor (#batch, time1, d_model). torch.Tensor: Cache tensor (1, head, cache_t + time1, d_k * 2) where `cache_t == chunk_size * num_decoding_left_chunks` and `head * d_k == size` """ q, k, v = self.forward_qkv(query, key, value) if cache.size(0) > 0: key_cache, value_cache = torch.split(cache, cache.size(-1) // 2, dim=-1) k = torch.cat([key_cache, k], dim=2) v = torch.cat([value_cache, v], dim=2) new_cache = torch.cat((k, v), dim=-1) scores = torch.matmul(q, k.transpose(-2, -1)) / math.sqrt(self.d_k) return self.forward_attention(v, scores, mask), new_cache class RelPositionMultiHeadedAttention(MultiHeadedAttention): """Multi-Head Attention layer with relative position encoding. Paper: https://arxiv.org/abs/1901.02860 Args: n_head (int): The number of heads. n_feat (int): The number of features. dropout_rate (float): Dropout rate. """ def __init__(self, n_head: int, n_feat: int, dropout_rate: float, key_bias: bool = True, dtype=None, device=None, operations=None): """Construct an RelPositionMultiHeadedAttention object.""" super().__init__(n_head, n_feat, dropout_rate, key_bias, dtype=dtype, device=device, operations=operations) # linear transformation for positional encoding self.linear_pos = operations.Linear(n_feat, n_feat, bias=False, dtype=dtype, device=device) # these two learnable bias are used in matrix c and matrix d # as described in https://arxiv.org/abs/1901.02860 Section 3.3 self.pos_bias_u = nn.Parameter(torch.empty(self.h, self.d_k, dtype=dtype, device=device)) self.pos_bias_v = nn.Parameter(torch.empty(self.h, self.d_k, dtype=dtype, device=device)) # torch.nn.init.xavier_uniform_(self.pos_bias_u) # torch.nn.init.xavier_uniform_(self.pos_bias_v) def rel_shift(self, x: torch.Tensor) -> torch.Tensor: """Compute relative positional encoding. Args: x (torch.Tensor): Input tensor (batch, head, time1, 2*time1-1). time1 means the length of query vector. Returns: torch.Tensor: Output tensor. """ zero_pad = torch.zeros((x.size()[0], x.size()[1], x.size()[2], 1), device=x.device, dtype=x.dtype) x_padded = torch.cat([zero_pad, x], dim=-1) x_padded = x_padded.view(x.size()[0], x.size()[1], x.size(3) + 1, x.size(2)) x = x_padded[:, :, 1:].view_as(x)[ :, :, :, : x.size(-1) // 2 + 1 ] # only keep the positions from 0 to time2 return x def forward( self, query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, mask: torch.Tensor = torch.ones((0, 0, 0), dtype=torch.bool), pos_emb: torch.Tensor = torch.empty(0), cache: torch.Tensor = torch.zeros((0, 0, 0, 0)) ) -> Tuple[torch.Tensor, torch.Tensor]: """Compute 'Scaled Dot Product Attention' with rel. positional encoding. Args: query (torch.Tensor): Query tensor (#batch, time1, size). key (torch.Tensor): Key tensor (#batch, time2, size). value (torch.Tensor): Value tensor (#batch, time2, size). mask (torch.Tensor): Mask tensor (#batch, 1, time2) or (#batch, time1, time2), (0, 0, 0) means fake mask. pos_emb (torch.Tensor): Positional embedding tensor (#batch, time2, size). cache (torch.Tensor): Cache tensor (1, head, cache_t, d_k * 2), where `cache_t == chunk_size * num_decoding_left_chunks` and `head * d_k == size` Returns: torch.Tensor: Output tensor (#batch, time1, d_model). torch.Tensor: Cache tensor (1, head, cache_t + time1, d_k * 2) where `cache_t == chunk_size * num_decoding_left_chunks` and `head * d_k == size` """ q, k, v = self.forward_qkv(query, key, value) q = q.transpose(1, 2) # (batch, time1, head, d_k) if cache.size(0) > 0: key_cache, value_cache = torch.split(cache, cache.size(-1) // 2, dim=-1) k = torch.cat([key_cache, k], dim=2) v = torch.cat([value_cache, v], dim=2) # NOTE(xcsong): We do cache slicing in encoder.forward_chunk, since it's # non-trivial to calculate `next_cache_start` here. new_cache = torch.cat((k, v), dim=-1) n_batch_pos = pos_emb.size(0) p = self.linear_pos(pos_emb).view(n_batch_pos, -1, self.h, self.d_k) p = p.transpose(1, 2) # (batch, head, time1, d_k) # (batch, head, time1, d_k) q_with_bias_u = (q + comfy.model_management.cast_to(self.pos_bias_u, dtype=q.dtype, device=q.device)).transpose(1, 2) # (batch, head, time1, d_k) q_with_bias_v = (q + comfy.model_management.cast_to(self.pos_bias_v, dtype=q.dtype, device=q.device)).transpose(1, 2) # compute attention score # first compute matrix a and matrix c # as described in https://arxiv.org/abs/1901.02860 Section 3.3 # (batch, head, time1, time2) matrix_ac = torch.matmul(q_with_bias_u, k.transpose(-2, -1)) # compute matrix b and matrix d # (batch, head, time1, time2) matrix_bd = torch.matmul(q_with_bias_v, p.transpose(-2, -1)) # NOTE(Xiang Lyu): Keep rel_shift since espnet rel_pos_emb is used if matrix_ac.shape != matrix_bd.shape: matrix_bd = self.rel_shift(matrix_bd) scores = (matrix_ac + matrix_bd) / math.sqrt( self.d_k) # (batch, head, time1, time2) return self.forward_attention(v, scores, mask), new_cache def subsequent_mask( size: int, device: torch.device = torch.device("cpu"), ) -> torch.Tensor: """Create mask for subsequent steps (size, size). This mask is used only in decoder which works in an auto-regressive mode. This means the current step could only do attention with its left steps. In encoder, fully attention is used when streaming is not necessary and the sequence is not long. In this case, no attention mask is needed. When streaming is need, chunk-based attention is used in encoder. See subsequent_chunk_mask for the chunk-based attention mask. Args: size (int): size of mask str device (str): "cpu" or "cuda" or torch.Tensor.device dtype (torch.device): result dtype Returns: torch.Tensor: mask Examples: >>> subsequent_mask(3) [[1, 0, 0], [1, 1, 0], [1, 1, 1]] """ arange = torch.arange(size, device=device) mask = arange.expand(size, size) arange = arange.unsqueeze(-1) mask = mask <= arange return mask def subsequent_chunk_mask( size: int, chunk_size: int, num_left_chunks: int = -1, device: torch.device = torch.device("cpu"), ) -> torch.Tensor: """Create mask for subsequent steps (size, size) with chunk size, this is for streaming encoder Args: size (int): size of mask chunk_size (int): size of chunk num_left_chunks (int): number of left chunks <0: use full chunk >=0: use num_left_chunks device (torch.device): "cpu" or "cuda" or torch.Tensor.device Returns: torch.Tensor: mask Examples: >>> subsequent_chunk_mask(4, 2) [[1, 1, 0, 0], [1, 1, 0, 0], [1, 1, 1, 1], [1, 1, 1, 1]] """ ret = torch.zeros(size, size, device=device, dtype=torch.bool) for i in range(size): if num_left_chunks < 0: start = 0 else: start = max((i // chunk_size - num_left_chunks) * chunk_size, 0) ending = min((i // chunk_size + 1) * chunk_size, size) ret[i, start:ending] = True return ret def add_optional_chunk_mask(xs: torch.Tensor, masks: torch.Tensor, use_dynamic_chunk: bool, use_dynamic_left_chunk: bool, decoding_chunk_size: int, static_chunk_size: int, num_decoding_left_chunks: int, enable_full_context: bool = True): """ Apply optional mask for encoder. Args: xs (torch.Tensor): padded input, (B, L, D), L for max length mask (torch.Tensor): mask for xs, (B, 1, L) use_dynamic_chunk (bool): whether to use dynamic chunk or not use_dynamic_left_chunk (bool): whether to use dynamic left chunk for training. decoding_chunk_size (int): decoding chunk size for dynamic chunk, it's 0: default for training, use random dynamic chunk. <0: for decoding, use full chunk. >0: for decoding, use fixed chunk size as set. static_chunk_size (int): chunk size for static chunk training/decoding if it's greater than 0, if use_dynamic_chunk is true, this parameter will be ignored num_decoding_left_chunks: number of left chunks, this is for decoding, the chunk size is decoding_chunk_size. >=0: use num_decoding_left_chunks <0: use all left chunks enable_full_context (bool): True: chunk size is either [1, 25] or full context(max_len) False: chunk size ~ U[1, 25] Returns: torch.Tensor: chunk mask of the input xs. """ # Whether to use chunk mask or not if use_dynamic_chunk: max_len = xs.size(1) if decoding_chunk_size < 0: chunk_size = max_len num_left_chunks = -1 elif decoding_chunk_size > 0: chunk_size = decoding_chunk_size num_left_chunks = num_decoding_left_chunks else: # chunk size is either [1, 25] or full context(max_len). # Since we use 4 times subsampling and allow up to 1s(100 frames) # delay, the maximum frame is 100 / 4 = 25. chunk_size = torch.randint(1, max_len, (1, )).item() num_left_chunks = -1 if chunk_size > max_len // 2 and enable_full_context: chunk_size = max_len else: chunk_size = chunk_size % 25 + 1 if use_dynamic_left_chunk: max_left_chunks = (max_len - 1) // chunk_size num_left_chunks = torch.randint(0, max_left_chunks, (1, )).item() chunk_masks = subsequent_chunk_mask(xs.size(1), chunk_size, num_left_chunks, xs.device) # (L, L) chunk_masks = chunk_masks.unsqueeze(0) # (1, L, L) chunk_masks = masks & chunk_masks # (B, L, L) elif static_chunk_size > 0: num_left_chunks = num_decoding_left_chunks chunk_masks = subsequent_chunk_mask(xs.size(1), static_chunk_size, num_left_chunks, xs.device) # (L, L) chunk_masks = chunk_masks.unsqueeze(0) # (1, L, L) chunk_masks = masks & chunk_masks # (B, L, L) else: chunk_masks = masks return chunk_masks class ConformerEncoderLayer(nn.Module): """Encoder layer module. Args: size (int): Input dimension. self_attn (torch.nn.Module): Self-attention module instance. `MultiHeadedAttention` or `RelPositionMultiHeadedAttention` instance can be used as the argument. feed_forward (torch.nn.Module): Feed-forward module instance. `PositionwiseFeedForward` instance can be used as the argument. feed_forward_macaron (torch.nn.Module): Additional feed-forward module instance. `PositionwiseFeedForward` instance can be used as the argument. conv_module (torch.nn.Module): Convolution module instance. `ConvlutionModule` instance can be used as the argument. dropout_rate (float): Dropout rate. normalize_before (bool): True: use layer_norm before each sub-block. False: use layer_norm after each sub-block. """ def __init__( self, size: int, self_attn: torch.nn.Module, feed_forward: Optional[nn.Module] = None, feed_forward_macaron: Optional[nn.Module] = None, conv_module: Optional[nn.Module] = None, dropout_rate: float = 0.1, normalize_before: bool = True, dtype=None, device=None, operations=None ): """Construct an EncoderLayer object.""" super().__init__() self.self_attn = self_attn self.feed_forward = feed_forward self.feed_forward_macaron = feed_forward_macaron self.conv_module = conv_module self.norm_ff = operations.LayerNorm(size, eps=1e-5, dtype=dtype, device=device) # for the FNN module self.norm_mha = operations.LayerNorm(size, eps=1e-5, dtype=dtype, device=device) # for the MHA module if feed_forward_macaron is not None: self.norm_ff_macaron = operations.LayerNorm(size, eps=1e-5, dtype=dtype, device=device) self.ff_scale = 0.5 else: self.ff_scale = 1.0 if self.conv_module is not None: self.norm_conv = operations.LayerNorm(size, eps=1e-5, dtype=dtype, device=device) # for the CNN module self.norm_final = operations.LayerNorm( size, eps=1e-5, dtype=dtype, device=device) # for the final output of the block self.dropout = nn.Dropout(dropout_rate) self.size = size self.normalize_before = normalize_before def forward( self, x: torch.Tensor, mask: torch.Tensor, pos_emb: torch.Tensor, mask_pad: torch.Tensor = torch.ones((0, 0, 0), dtype=torch.bool), att_cache: torch.Tensor = torch.zeros((0, 0, 0, 0)), cnn_cache: torch.Tensor = torch.zeros((0, 0, 0, 0)), ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor]: """Compute encoded features. Args: x (torch.Tensor): (#batch, time, size) mask (torch.Tensor): Mask tensor for the input (#batch, time,time), (0, 0, 0) means fake mask. pos_emb (torch.Tensor): positional encoding, must not be None for ConformerEncoderLayer. mask_pad (torch.Tensor): batch padding mask used for conv module. (#batch, 1,time), (0, 0, 0) means fake mask. att_cache (torch.Tensor): Cache tensor of the KEY & VALUE (#batch=1, head, cache_t1, d_k * 2), head * d_k == size. cnn_cache (torch.Tensor): Convolution cache in conformer layer (#batch=1, size, cache_t2) Returns: torch.Tensor: Output tensor (#batch, time, size). torch.Tensor: Mask tensor (#batch, time, time). torch.Tensor: att_cache tensor, (#batch=1, head, cache_t1 + time, d_k * 2). torch.Tensor: cnn_cahce tensor (#batch, size, cache_t2). """ # whether to use macaron style if self.feed_forward_macaron is not None: residual = x if self.normalize_before: x = self.norm_ff_macaron(x) x = residual + self.ff_scale * self.dropout( self.feed_forward_macaron(x)) if not self.normalize_before: x = self.norm_ff_macaron(x) # multi-headed self-attention module residual = x if self.normalize_before: x = self.norm_mha(x) x_att, new_att_cache = self.self_attn(x, x, x, mask, pos_emb, att_cache) x = residual + self.dropout(x_att) if not self.normalize_before: x = self.norm_mha(x) # convolution module # Fake new cnn cache here, and then change it in conv_module new_cnn_cache = torch.zeros((0, 0, 0), dtype=x.dtype, device=x.device) if self.conv_module is not None: residual = x if self.normalize_before: x = self.norm_conv(x) x, new_cnn_cache = self.conv_module(x, mask_pad, cnn_cache) x = residual + self.dropout(x) if not self.normalize_before: x = self.norm_conv(x) # feed forward module residual = x if self.normalize_before: x = self.norm_ff(x) x = residual + self.ff_scale * self.dropout(self.feed_forward(x)) if not self.normalize_before: x = self.norm_ff(x) if self.conv_module is not None: x = self.norm_final(x) return x, mask, new_att_cache, new_cnn_cache class EspnetRelPositionalEncoding(torch.nn.Module): """Relative positional encoding module (new implementation). Details can be found in https://github.com/espnet/espnet/pull/2816. See : Appendix B in https://arxiv.org/abs/1901.02860 Args: d_model (int): Embedding dimension. dropout_rate (float): Dropout rate. max_len (int): Maximum input length. """ def __init__(self, d_model: int, dropout_rate: float, max_len: int = 5000): """Construct an PositionalEncoding object.""" super(EspnetRelPositionalEncoding, self).__init__() self.d_model = d_model self.xscale = math.sqrt(self.d_model) self.dropout = torch.nn.Dropout(p=dropout_rate) self.pe = None self.extend_pe(torch.tensor(0.0).expand(1, max_len)) def extend_pe(self, x: torch.Tensor): """Reset the positional encodings.""" if self.pe is not None: # self.pe contains both positive and negative parts # the length of self.pe is 2 * input_len - 1 if self.pe.size(1) >= x.size(1) * 2 - 1: if self.pe.dtype != x.dtype or self.pe.device != x.device: self.pe = self.pe.to(dtype=x.dtype, device=x.device) return # Suppose `i` means to the position of query vecotr and `j` means the # position of key vector. We use position relative positions when keys # are to the left (i>j) and negative relative positions otherwise (i<j). pe_positive = torch.zeros(x.size(1), self.d_model) pe_negative = torch.zeros(x.size(1), self.d_model) position = torch.arange(0, x.size(1), dtype=torch.float32).unsqueeze(1) div_term = torch.exp( torch.arange(0, self.d_model, 2, dtype=torch.float32) * -(math.log(10000.0) / self.d_model) ) pe_positive[:, 0::2] = torch.sin(position * div_term) pe_positive[:, 1::2] = torch.cos(position * div_term) pe_negative[:, 0::2] = torch.sin(-1 * position * div_term) pe_negative[:, 1::2] = torch.cos(-1 * position * div_term) # Reserve the order of positive indices and concat both positive and # negative indices. This is used to support the shifting trick # as in https://arxiv.org/abs/1901.02860 pe_positive = torch.flip(pe_positive, [0]).unsqueeze(0) pe_negative = pe_negative[1:].unsqueeze(0) pe = torch.cat([pe_positive, pe_negative], dim=1) self.pe = pe.to(device=x.device, dtype=x.dtype) def forward(self, x: torch.Tensor, offset: Union[int, torch.Tensor] = 0) \ -> Tuple[torch.Tensor, torch.Tensor]: """Add positional encoding. Args: x (torch.Tensor): Input tensor (batch, time, `*`). Returns: torch.Tensor: Encoded tensor (batch, time, `*`). """ self.extend_pe(x) x = x * self.xscale pos_emb = self.position_encoding(size=x.size(1), offset=offset) return self.dropout(x), self.dropout(pos_emb) def position_encoding(self, offset: Union[int, torch.Tensor], size: int) -> torch.Tensor: """ For getting encoding in a streaming fashion Attention!!!!! we apply dropout only once at the whole utterance level in a none streaming way, but will call this function several times with increasing input size in a streaming scenario, so the dropout will be applied several times. Args: offset (int or torch.tensor): start offset size (int): required size of position encoding Returns: torch.Tensor: Corresponding encoding """ pos_emb = self.pe[ :, self.pe.size(1) // 2 - size + 1: self.pe.size(1) // 2 + size, ] return pos_emb class LinearEmbed(torch.nn.Module): """Linear transform the input without subsampling Args: idim (int): Input dimension. odim (int): Output dimension. dropout_rate (float): Dropout rate. """ def __init__(self, idim: int, odim: int, dropout_rate: float, pos_enc_class: torch.nn.Module, dtype=None, device=None, operations=None): """Construct an linear object.""" super().__init__() self.out = torch.nn.Sequential( operations.Linear(idim, odim, dtype=dtype, device=device), operations.LayerNorm(odim, eps=1e-5, dtype=dtype, device=device), torch.nn.Dropout(dropout_rate), ) self.pos_enc = pos_enc_class #rel_pos_espnet def position_encoding(self, offset: Union[int, torch.Tensor], size: int) -> torch.Tensor: return self.pos_enc.position_encoding(offset, size) def forward( self, x: torch.Tensor, offset: Union[int, torch.Tensor] = 0 ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: """Input x. Args: x (torch.Tensor): Input tensor (#batch, time, idim). x_mask (torch.Tensor): Input mask (#batch, 1, time). Returns: torch.Tensor: linear input tensor (#batch, time', odim), where time' = time . torch.Tensor: linear input mask (#batch, 1, time'), where time' = time . """ x = self.out(x) x, pos_emb = self.pos_enc(x, offset) return x, pos_emb ATTENTION_CLASSES = { "selfattn": MultiHeadedAttention, "rel_selfattn": RelPositionMultiHeadedAttention, } ACTIVATION_CLASSES = { "hardtanh": torch.nn.Hardtanh, "tanh": torch.nn.Tanh, "relu": torch.nn.ReLU, "selu": torch.nn.SELU, "swish": getattr(torch.nn, "SiLU", Swish), "gelu": torch.nn.GELU, } def make_pad_mask(lengths: torch.Tensor, max_len: int = 0) -> torch.Tensor: """Make mask tensor containing indices of padded part. See description of make_non_pad_mask. Args: lengths (torch.Tensor): Batch of lengths (B,). Returns: torch.Tensor: Mask tensor containing indices of padded part. Examples: >>> lengths = [5, 3, 2] >>> make_pad_mask(lengths) masks = [[0, 0, 0, 0 ,0], [0, 0, 0, 1, 1], [0, 0, 1, 1, 1]] """ batch_size = lengths.size(0) max_len = max_len if max_len > 0 else lengths.max().item() seq_range = torch.arange(0, max_len, dtype=torch.int64, device=lengths.device) seq_range_expand = seq_range.unsqueeze(0).expand(batch_size, max_len) seq_length_expand = lengths.unsqueeze(-1) mask = seq_range_expand >= seq_length_expand return mask #https://github.com/FunAudioLLM/CosyVoice/blob/main/examples/magicdata-read/cosyvoice/conf/cosyvoice.yaml class ConformerEncoder(torch.nn.Module): """Conformer encoder module.""" def __init__( self, input_size: int, output_size: int = 1024, attention_heads: int = 16, linear_units: int = 4096, num_blocks: int = 6, dropout_rate: float = 0.1, positional_dropout_rate: float = 0.1, attention_dropout_rate: float = 0.0, input_layer: str = 'linear', pos_enc_layer_type: str = 'rel_pos_espnet', normalize_before: bool = True, static_chunk_size: int = 1, # 1: causal_mask; 0: full_mask use_dynamic_chunk: bool = False, use_dynamic_left_chunk: bool = False, positionwise_conv_kernel_size: int = 1, macaron_style: bool =False, selfattention_layer_type: str = "rel_selfattn", activation_type: str = "swish", use_cnn_module: bool = False, cnn_module_kernel: int = 15, causal: bool = False, cnn_module_norm: str = "batch_norm", key_bias: bool = True, dtype=None, device=None, operations=None ): """Construct ConformerEncoder Args: input_size to use_dynamic_chunk, see in BaseEncoder positionwise_conv_kernel_size (int): Kernel size of positionwise conv1d layer. macaron_style (bool): Whether to use macaron style for positionwise layer. selfattention_layer_type (str): Encoder attention layer type, the parameter has no effect now, it's just for configure compatibility. #'rel_selfattn' activation_type (str): Encoder activation function type. use_cnn_module (bool): Whether to use convolution module. cnn_module_kernel (int): Kernel size of convolution module. causal (bool): whether to use causal convolution or not. key_bias: whether use bias in attention.linear_k, False for whisper models. """ super().__init__() self.output_size = output_size self.embed = LinearEmbed(input_size, output_size, dropout_rate, EspnetRelPositionalEncoding(output_size, positional_dropout_rate), dtype=dtype, device=device, operations=operations) self.normalize_before = normalize_before self.after_norm = operations.LayerNorm(output_size, eps=1e-5, dtype=dtype, device=device) self.use_dynamic_chunk = use_dynamic_chunk self.static_chunk_size = static_chunk_size self.use_dynamic_chunk = use_dynamic_chunk self.use_dynamic_left_chunk = use_dynamic_left_chunk activation = ACTIVATION_CLASSES[activation_type]() # self-attention module definition encoder_selfattn_layer_args = ( attention_heads, output_size, attention_dropout_rate, key_bias, ) # feed-forward module definition positionwise_layer_args = ( output_size, linear_units, dropout_rate, activation, ) # convolution module definition convolution_layer_args = (output_size, cnn_module_kernel, activation, cnn_module_norm, causal) self.encoders = torch.nn.ModuleList([ ConformerEncoderLayer( output_size, RelPositionMultiHeadedAttention( *encoder_selfattn_layer_args, dtype=dtype, device=device, operations=operations), PositionwiseFeedForward(*positionwise_layer_args, dtype=dtype, device=device, operations=operations), PositionwiseFeedForward( *positionwise_layer_args, dtype=dtype, device=device, operations=operations) if macaron_style else None, ConvolutionModule( *convolution_layer_args, dtype=dtype, device=device, operations=operations) if use_cnn_module else None, dropout_rate, normalize_before, dtype=dtype, device=device, operations=operations ) for _ in range(num_blocks) ]) def forward_layers(self, xs: torch.Tensor, chunk_masks: torch.Tensor, pos_emb: torch.Tensor, mask_pad: torch.Tensor) -> torch.Tensor: for layer in self.encoders: xs, chunk_masks, _, _ = layer(xs, chunk_masks, pos_emb, mask_pad) return xs def forward( self, xs: torch.Tensor, pad_mask: torch.Tensor, decoding_chunk_size: int = 0, num_decoding_left_chunks: int = -1, ) -> Tuple[torch.Tensor, torch.Tensor]: """Embed positions in tensor. Args: xs: padded input tensor (B, T, D) xs_lens: input length (B) decoding_chunk_size: decoding chunk size for dynamic chunk 0: default for training, use random dynamic chunk. <0: for decoding, use full chunk. >0: for decoding, use fixed chunk size as set. num_decoding_left_chunks: number of left chunks, this is for decoding, the chunk size is decoding_chunk_size. >=0: use num_decoding_left_chunks <0: use all left chunks Returns: encoder output tensor xs, and subsampled masks xs: padded output tensor (B, T' ~= T/subsample_rate, D) masks: torch.Tensor batch padding mask after subsample (B, 1, T' ~= T/subsample_rate) NOTE(xcsong): We pass the `__call__` method of the modules instead of `forward` to the checkpointing API because `__call__` attaches all the hooks of the module. https://discuss.pytorch.org/t/any-different-between-model-input-and-model-forward-input/3690/2 """ masks = None if pad_mask is not None: masks = pad_mask.to(torch.bool).unsqueeze(1) # (B, 1, T) xs, pos_emb = self.embed(xs) mask_pad = masks # (B, 1, T/subsample_rate) chunk_masks = add_optional_chunk_mask(xs, masks, self.use_dynamic_chunk, self.use_dynamic_left_chunk, decoding_chunk_size, self.static_chunk_size, num_decoding_left_chunks) xs = self.forward_layers(xs, chunk_masks, pos_emb, mask_pad) if self.normalize_before: xs = self.after_norm(xs) # Here we assume the mask is not changed in encoder layers, so just # return the masks before encoder layers, and the masks will be used # for cross attention with decoder later return xs, masks
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/ldm/ace/lyric_encoder.py", "license": "GNU General Public License v3.0", "lines": 931, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy/ldm/ace/model.py
# Original from: https://github.com/ace-step/ACE-Step/blob/main/models/ace_step_transformer.py # Copyright 2024 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import Optional, List, Union import torch from torch import nn import comfy.model_management import comfy.patcher_extension from comfy.ldm.lightricks.model import TimestepEmbedding, Timesteps from .attention import LinearTransformerBlock, t2i_modulate from .lyric_encoder import ConformerEncoder as LyricEncoder def cross_norm(hidden_states, controlnet_input): # input N x T x c mean_hidden_states, std_hidden_states = hidden_states.mean(dim=(1,2), keepdim=True), hidden_states.std(dim=(1,2), keepdim=True) mean_controlnet_input, std_controlnet_input = controlnet_input.mean(dim=(1,2), keepdim=True), controlnet_input.std(dim=(1,2), keepdim=True) controlnet_input = (controlnet_input - mean_controlnet_input) * (std_hidden_states / (std_controlnet_input + 1e-12)) + mean_hidden_states return controlnet_input # Copied from transformers.models.mixtral.modeling_mixtral.MixtralRotaryEmbedding with Mixtral->Qwen2 class Qwen2RotaryEmbedding(nn.Module): def __init__(self, dim, max_position_embeddings=2048, base=10000, dtype=None, device=None): super().__init__() self.dim = dim self.max_position_embeddings = max_position_embeddings self.base = base inv_freq = 1.0 / (self.base ** (torch.arange(0, self.dim, 2, dtype=torch.int64, device=device).float() / self.dim)) self.register_buffer("inv_freq", inv_freq, persistent=False) # Build here to make `torch.jit.trace` work. self._set_cos_sin_cache( seq_len=max_position_embeddings, device=self.inv_freq.device, dtype=torch.float32 ) def _set_cos_sin_cache(self, seq_len, device, dtype): self.max_seq_len_cached = seq_len t = torch.arange(self.max_seq_len_cached, device=device, dtype=torch.int64).type_as(self.inv_freq) freqs = torch.outer(t, self.inv_freq) # Different from paper, but it uses a different permutation in order to obtain the same calculation emb = torch.cat((freqs, freqs), dim=-1) self.register_buffer("cos_cached", emb.cos().to(dtype), persistent=False) self.register_buffer("sin_cached", emb.sin().to(dtype), persistent=False) def forward(self, x, seq_len=None): # x: [bs, num_attention_heads, seq_len, head_size] if seq_len > self.max_seq_len_cached: self._set_cos_sin_cache(seq_len=seq_len, device=x.device, dtype=x.dtype) return ( self.cos_cached[:seq_len].to(dtype=x.dtype), self.sin_cached[:seq_len].to(dtype=x.dtype), ) class T2IFinalLayer(nn.Module): """ The final layer of Sana. """ def __init__(self, hidden_size, patch_size=[16, 1], out_channels=256, dtype=None, device=None, operations=None): super().__init__() self.norm_final = operations.RMSNorm(hidden_size, elementwise_affine=False, eps=1e-6, dtype=dtype, device=device) self.linear = operations.Linear(hidden_size, patch_size[0] * patch_size[1] * out_channels, bias=True, dtype=dtype, device=device) self.scale_shift_table = nn.Parameter(torch.empty(2, hidden_size, dtype=dtype, device=device)) self.out_channels = out_channels self.patch_size = patch_size def unpatchfy( self, hidden_states: torch.Tensor, width: int, ): # 4 unpatchify new_height, new_width = 1, hidden_states.size(1) hidden_states = hidden_states.reshape( shape=(hidden_states.shape[0], new_height, new_width, self.patch_size[0], self.patch_size[1], self.out_channels) ).contiguous() hidden_states = torch.einsum("nhwpqc->nchpwq", hidden_states) output = hidden_states.reshape( shape=(hidden_states.shape[0], self.out_channels, new_height * self.patch_size[0], new_width * self.patch_size[1]) ).contiguous() if width > new_width: output = torch.nn.functional.pad(output, (0, width - new_width, 0, 0), 'constant', 0) elif width < new_width: output = output[:, :, :, :width] return output def forward(self, x, t, output_length): shift, scale = (comfy.model_management.cast_to(self.scale_shift_table[None], device=t.device, dtype=t.dtype) + t[:, None]).chunk(2, dim=1) x = t2i_modulate(self.norm_final(x), shift, scale) x = self.linear(x) # unpatchify output = self.unpatchfy(x, output_length) return output class PatchEmbed(nn.Module): """2D Image to Patch Embedding""" def __init__( self, height=16, width=4096, patch_size=(16, 1), in_channels=8, embed_dim=1152, bias=True, dtype=None, device=None, operations=None ): super().__init__() patch_size_h, patch_size_w = patch_size self.early_conv_layers = nn.Sequential( operations.Conv2d(in_channels, in_channels*256, kernel_size=patch_size, stride=patch_size, padding=0, bias=bias, dtype=dtype, device=device), operations.GroupNorm(num_groups=32, num_channels=in_channels*256, eps=1e-6, affine=True, dtype=dtype, device=device), operations.Conv2d(in_channels*256, embed_dim, kernel_size=1, stride=1, padding=0, bias=bias, dtype=dtype, device=device) ) self.patch_size = patch_size self.height, self.width = height // patch_size_h, width // patch_size_w self.base_size = self.width def forward(self, latent): # early convolutions, N x C x H x W -> N x 256 * sqrt(patch_size) x H/patch_size x W/patch_size latent = self.early_conv_layers(latent) latent = latent.flatten(2).transpose(1, 2) # BCHW -> BNC return latent class ACEStepTransformer2DModel(nn.Module): # _supports_gradient_checkpointing = True def __init__( self, in_channels: Optional[int] = 8, num_layers: int = 28, inner_dim: int = 1536, attention_head_dim: int = 64, num_attention_heads: int = 24, mlp_ratio: float = 4.0, out_channels: int = 8, max_position: int = 32768, rope_theta: float = 1000000.0, speaker_embedding_dim: int = 512, text_embedding_dim: int = 768, ssl_encoder_depths: List[int] = [9, 9], ssl_names: List[str] = ["mert", "m-hubert"], ssl_latent_dims: List[int] = [1024, 768], lyric_encoder_vocab_size: int = 6681, lyric_hidden_size: int = 1024, patch_size: List[int] = [16, 1], max_height: int = 16, max_width: int = 4096, audio_model=None, dtype=None, device=None, operations=None ): super().__init__() self.dtype = dtype self.num_attention_heads = num_attention_heads self.attention_head_dim = attention_head_dim inner_dim = num_attention_heads * attention_head_dim self.inner_dim = inner_dim self.out_channels = out_channels self.max_position = max_position self.patch_size = patch_size self.rope_theta = rope_theta self.rotary_emb = Qwen2RotaryEmbedding( dim=self.attention_head_dim, max_position_embeddings=self.max_position, base=self.rope_theta, dtype=dtype, device=device, ) # 2. Define input layers self.in_channels = in_channels self.num_layers = num_layers # 3. Define transformers blocks self.transformer_blocks = nn.ModuleList( [ LinearTransformerBlock( dim=self.inner_dim, num_attention_heads=self.num_attention_heads, attention_head_dim=attention_head_dim, mlp_ratio=mlp_ratio, add_cross_attention=True, add_cross_attention_dim=self.inner_dim, dtype=dtype, device=device, operations=operations, ) for i in range(self.num_layers) ] ) self.time_proj = Timesteps(num_channels=256, flip_sin_to_cos=True, downscale_freq_shift=0) self.timestep_embedder = TimestepEmbedding(in_channels=256, time_embed_dim=self.inner_dim, dtype=dtype, device=device, operations=operations) self.t_block = nn.Sequential(nn.SiLU(), operations.Linear(self.inner_dim, 6 * self.inner_dim, bias=True, dtype=dtype, device=device)) # speaker self.speaker_embedder = operations.Linear(speaker_embedding_dim, self.inner_dim, dtype=dtype, device=device) # genre self.genre_embedder = operations.Linear(text_embedding_dim, self.inner_dim, dtype=dtype, device=device) # lyric self.lyric_embs = operations.Embedding(lyric_encoder_vocab_size, lyric_hidden_size, dtype=dtype, device=device) self.lyric_encoder = LyricEncoder(input_size=lyric_hidden_size, static_chunk_size=0, dtype=dtype, device=device, operations=operations) self.lyric_proj = operations.Linear(lyric_hidden_size, self.inner_dim, dtype=dtype, device=device) projector_dim = 2 * self.inner_dim self.projectors = nn.ModuleList([ nn.Sequential( operations.Linear(self.inner_dim, projector_dim, dtype=dtype, device=device), nn.SiLU(), operations.Linear(projector_dim, projector_dim, dtype=dtype, device=device), nn.SiLU(), operations.Linear(projector_dim, ssl_dim, dtype=dtype, device=device), ) for ssl_dim in ssl_latent_dims ]) self.proj_in = PatchEmbed( height=max_height, width=max_width, patch_size=patch_size, embed_dim=self.inner_dim, bias=True, dtype=dtype, device=device, operations=operations, ) self.final_layer = T2IFinalLayer(self.inner_dim, patch_size=patch_size, out_channels=out_channels, dtype=dtype, device=device, operations=operations) def forward_lyric_encoder( self, lyric_token_idx: Optional[torch.LongTensor] = None, lyric_mask: Optional[torch.LongTensor] = None, out_dtype=None, ): # N x T x D lyric_embs = self.lyric_embs(lyric_token_idx, out_dtype=out_dtype) prompt_prenet_out, _mask = self.lyric_encoder(lyric_embs, lyric_mask, decoding_chunk_size=1, num_decoding_left_chunks=-1) prompt_prenet_out = self.lyric_proj(prompt_prenet_out) return prompt_prenet_out def encode( self, encoder_text_hidden_states: Optional[torch.Tensor] = None, text_attention_mask: Optional[torch.LongTensor] = None, speaker_embeds: Optional[torch.FloatTensor] = None, lyric_token_idx: Optional[torch.LongTensor] = None, lyric_mask: Optional[torch.LongTensor] = None, lyrics_strength=1.0, ): bs = encoder_text_hidden_states.shape[0] device = encoder_text_hidden_states.device # speaker embedding encoder_spk_hidden_states = self.speaker_embedder(speaker_embeds).unsqueeze(1) # genre embedding encoder_text_hidden_states = self.genre_embedder(encoder_text_hidden_states) # lyric encoder_lyric_hidden_states = self.forward_lyric_encoder( lyric_token_idx=lyric_token_idx, lyric_mask=lyric_mask, out_dtype=encoder_text_hidden_states.dtype, ) encoder_lyric_hidden_states *= lyrics_strength encoder_hidden_states = torch.cat([encoder_spk_hidden_states, encoder_text_hidden_states, encoder_lyric_hidden_states], dim=1) encoder_hidden_mask = None if text_attention_mask is not None: speaker_mask = torch.ones(bs, 1, device=device) encoder_hidden_mask = torch.cat([speaker_mask, text_attention_mask, lyric_mask], dim=1) return encoder_hidden_states, encoder_hidden_mask def decode( self, hidden_states: torch.Tensor, attention_mask: torch.Tensor, encoder_hidden_states: torch.Tensor, encoder_hidden_mask: torch.Tensor, timestep: Optional[torch.Tensor], output_length: int = 0, block_controlnet_hidden_states: Optional[Union[List[torch.Tensor], torch.Tensor]] = None, controlnet_scale: Union[float, torch.Tensor] = 1.0, transformer_options={}, ): embedded_timestep = self.timestep_embedder(self.time_proj(timestep).to(dtype=hidden_states.dtype)) temb = self.t_block(embedded_timestep) hidden_states = self.proj_in(hidden_states) # controlnet logic if block_controlnet_hidden_states is not None: control_condi = cross_norm(hidden_states, block_controlnet_hidden_states) hidden_states = hidden_states + control_condi * controlnet_scale # inner_hidden_states = [] rotary_freqs_cis = self.rotary_emb(hidden_states, seq_len=hidden_states.shape[1]) encoder_rotary_freqs_cis = self.rotary_emb(encoder_hidden_states, seq_len=encoder_hidden_states.shape[1]) for index_block, block in enumerate(self.transformer_blocks): hidden_states = block( hidden_states=hidden_states, attention_mask=attention_mask, encoder_hidden_states=encoder_hidden_states, encoder_attention_mask=encoder_hidden_mask, rotary_freqs_cis=rotary_freqs_cis, rotary_freqs_cis_cross=encoder_rotary_freqs_cis, temb=temb, transformer_options=transformer_options, ) output = self.final_layer(hidden_states, embedded_timestep, output_length) return output def forward(self, x, timestep, attention_mask=None, context: Optional[torch.Tensor] = None, text_attention_mask: Optional[torch.LongTensor] = None, speaker_embeds: Optional[torch.FloatTensor] = None, lyric_token_idx: Optional[torch.LongTensor] = None, lyric_mask: Optional[torch.LongTensor] = None, block_controlnet_hidden_states: Optional[Union[List[torch.Tensor], torch.Tensor]] = None, controlnet_scale: Union[float, torch.Tensor] = 1.0, lyrics_strength=1.0, **kwargs ): return comfy.patcher_extension.WrapperExecutor.new_class_executor( self._forward, self, comfy.patcher_extension.get_all_wrappers(comfy.patcher_extension.WrappersMP.DIFFUSION_MODEL, kwargs.get("transformer_options", {})) ).execute(x, timestep, attention_mask, context, text_attention_mask, speaker_embeds, lyric_token_idx, lyric_mask, block_controlnet_hidden_states, controlnet_scale, lyrics_strength, **kwargs) def _forward( self, x, timestep, attention_mask=None, context: Optional[torch.Tensor] = None, text_attention_mask: Optional[torch.LongTensor] = None, speaker_embeds: Optional[torch.FloatTensor] = None, lyric_token_idx: Optional[torch.LongTensor] = None, lyric_mask: Optional[torch.LongTensor] = None, block_controlnet_hidden_states: Optional[Union[List[torch.Tensor], torch.Tensor]] = None, controlnet_scale: Union[float, torch.Tensor] = 1.0, lyrics_strength=1.0, **kwargs ): hidden_states = x encoder_text_hidden_states = context encoder_hidden_states, encoder_hidden_mask = self.encode( encoder_text_hidden_states=encoder_text_hidden_states, text_attention_mask=text_attention_mask, speaker_embeds=speaker_embeds, lyric_token_idx=lyric_token_idx, lyric_mask=lyric_mask, lyrics_strength=lyrics_strength, ) output_length = hidden_states.shape[-1] transformer_options = kwargs.get("transformer_options", {}) output = self.decode( hidden_states=hidden_states, attention_mask=attention_mask, encoder_hidden_states=encoder_hidden_states, encoder_hidden_mask=encoder_hidden_mask, timestep=timestep, output_length=output_length, block_controlnet_hidden_states=block_controlnet_hidden_states, controlnet_scale=controlnet_scale, transformer_options=transformer_options, ) return output
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/ldm/ace/model.py", "license": "GNU General Public License v3.0", "lines": 349, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
license
Comfy-Org/ComfyUI:comfy/ldm/ace/vae/autoencoder_dc.py
# Rewritten from diffusers import torch import torch.nn as nn import torch.nn.functional as F from typing import Tuple, Union import comfy.model_management import comfy.ops ops = comfy.ops.disable_weight_init class RMSNorm(ops.RMSNorm): def __init__(self, dim, eps=1e-5, elementwise_affine=True, bias=False): super().__init__(dim, eps=eps, elementwise_affine=elementwise_affine) if elementwise_affine: self.bias = nn.Parameter(torch.empty(dim)) if bias else None def forward(self, x): x = super().forward(x) if self.elementwise_affine: if self.bias is not None: x = x + comfy.model_management.cast_to(self.bias, dtype=x.dtype, device=x.device) return x def get_normalization(norm_type, num_features, num_groups=32, eps=1e-5): if norm_type == "batch_norm": return nn.BatchNorm2d(num_features) elif norm_type == "group_norm": return ops.GroupNorm(num_groups, num_features) elif norm_type == "layer_norm": return ops.LayerNorm(num_features) elif norm_type == "rms_norm": return RMSNorm(num_features, eps=eps, elementwise_affine=True, bias=True) else: raise ValueError(f"Unknown normalization type: {norm_type}") def get_activation(activation_type): if activation_type == "relu": return nn.ReLU() elif activation_type == "relu6": return nn.ReLU6() elif activation_type == "silu": return nn.SiLU() elif activation_type == "leaky_relu": return nn.LeakyReLU(0.2) else: raise ValueError(f"Unknown activation type: {activation_type}") class ResBlock(nn.Module): def __init__( self, in_channels: int, out_channels: int, norm_type: str = "batch_norm", act_fn: str = "relu6", ) -> None: super().__init__() self.norm_type = norm_type self.nonlinearity = get_activation(act_fn) if act_fn is not None else nn.Identity() self.conv1 = ops.Conv2d(in_channels, in_channels, 3, 1, 1) self.conv2 = ops.Conv2d(in_channels, out_channels, 3, 1, 1, bias=False) self.norm = get_normalization(norm_type, out_channels) def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: residual = hidden_states hidden_states = self.conv1(hidden_states) hidden_states = self.nonlinearity(hidden_states) hidden_states = self.conv2(hidden_states) if self.norm_type == "rms_norm": # move channel to the last dimension so we apply RMSnorm across channel dimension hidden_states = self.norm(hidden_states.movedim(1, -1)).movedim(-1, 1) else: hidden_states = self.norm(hidden_states) return hidden_states + residual class SanaMultiscaleAttentionProjection(nn.Module): def __init__( self, in_channels: int, num_attention_heads: int, kernel_size: int, ) -> None: super().__init__() channels = 3 * in_channels self.proj_in = ops.Conv2d( channels, channels, kernel_size, padding=kernel_size // 2, groups=channels, bias=False, ) self.proj_out = ops.Conv2d(channels, channels, 1, 1, 0, groups=3 * num_attention_heads, bias=False) def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: hidden_states = self.proj_in(hidden_states) hidden_states = self.proj_out(hidden_states) return hidden_states class SanaMultiscaleLinearAttention(nn.Module): def __init__( self, in_channels: int, out_channels: int, num_attention_heads: int = None, attention_head_dim: int = 8, mult: float = 1.0, norm_type: str = "batch_norm", kernel_sizes: tuple = (5,), eps: float = 1e-15, residual_connection: bool = False, ): super().__init__() self.eps = eps self.attention_head_dim = attention_head_dim self.norm_type = norm_type self.residual_connection = residual_connection num_attention_heads = ( int(in_channels // attention_head_dim * mult) if num_attention_heads is None else num_attention_heads ) inner_dim = num_attention_heads * attention_head_dim self.to_q = ops.Linear(in_channels, inner_dim, bias=False) self.to_k = ops.Linear(in_channels, inner_dim, bias=False) self.to_v = ops.Linear(in_channels, inner_dim, bias=False) self.to_qkv_multiscale = nn.ModuleList() for kernel_size in kernel_sizes: self.to_qkv_multiscale.append( SanaMultiscaleAttentionProjection(inner_dim, num_attention_heads, kernel_size) ) self.nonlinearity = nn.ReLU() self.to_out = ops.Linear(inner_dim * (1 + len(kernel_sizes)), out_channels, bias=False) self.norm_out = get_normalization(norm_type, out_channels) def apply_linear_attention(self, query, key, value): value = F.pad(value, (0, 0, 0, 1), mode="constant", value=1) scores = torch.matmul(value, key.transpose(-1, -2)) hidden_states = torch.matmul(scores, query) hidden_states = hidden_states.to(dtype=torch.float32) hidden_states = hidden_states[:, :, :-1] / (hidden_states[:, :, -1:] + self.eps) return hidden_states def apply_quadratic_attention(self, query, key, value): scores = torch.matmul(key.transpose(-1, -2), query) scores = scores.to(dtype=torch.float32) scores = scores / (torch.sum(scores, dim=2, keepdim=True) + self.eps) hidden_states = torch.matmul(value, scores.to(value.dtype)) return hidden_states def forward(self, hidden_states): height, width = hidden_states.shape[-2:] if height * width > self.attention_head_dim: use_linear_attention = True else: use_linear_attention = False residual = hidden_states batch_size, _, height, width = list(hidden_states.size()) original_dtype = hidden_states.dtype hidden_states = hidden_states.movedim(1, -1) query = self.to_q(hidden_states) key = self.to_k(hidden_states) value = self.to_v(hidden_states) hidden_states = torch.cat([query, key, value], dim=3) hidden_states = hidden_states.movedim(-1, 1) multi_scale_qkv = [hidden_states] for block in self.to_qkv_multiscale: multi_scale_qkv.append(block(hidden_states)) hidden_states = torch.cat(multi_scale_qkv, dim=1) if use_linear_attention: # for linear attention upcast hidden_states to float32 hidden_states = hidden_states.to(dtype=torch.float32) hidden_states = hidden_states.reshape(batch_size, -1, 3 * self.attention_head_dim, height * width) query, key, value = hidden_states.chunk(3, dim=2) query = self.nonlinearity(query) key = self.nonlinearity(key) if use_linear_attention: hidden_states = self.apply_linear_attention(query, key, value) hidden_states = hidden_states.to(dtype=original_dtype) else: hidden_states = self.apply_quadratic_attention(query, key, value) hidden_states = torch.reshape(hidden_states, (batch_size, -1, height, width)) hidden_states = self.to_out(hidden_states.movedim(1, -1)).movedim(-1, 1) if self.norm_type == "rms_norm": hidden_states = self.norm_out(hidden_states.movedim(1, -1)).movedim(-1, 1) else: hidden_states = self.norm_out(hidden_states) if self.residual_connection: hidden_states = hidden_states + residual return hidden_states class EfficientViTBlock(nn.Module): def __init__( self, in_channels: int, mult: float = 1.0, attention_head_dim: int = 32, qkv_multiscales: tuple = (5,), norm_type: str = "batch_norm", ) -> None: super().__init__() self.attn = SanaMultiscaleLinearAttention( in_channels=in_channels, out_channels=in_channels, mult=mult, attention_head_dim=attention_head_dim, norm_type=norm_type, kernel_sizes=qkv_multiscales, residual_connection=True, ) self.conv_out = GLUMBConv( in_channels=in_channels, out_channels=in_channels, norm_type="rms_norm", ) def forward(self, x: torch.Tensor) -> torch.Tensor: x = self.attn(x) x = self.conv_out(x) return x class GLUMBConv(nn.Module): def __init__( self, in_channels: int, out_channels: int, expand_ratio: float = 4, norm_type: str = None, residual_connection: bool = True, ) -> None: super().__init__() hidden_channels = int(expand_ratio * in_channels) self.norm_type = norm_type self.residual_connection = residual_connection self.nonlinearity = nn.SiLU() self.conv_inverted = ops.Conv2d(in_channels, hidden_channels * 2, 1, 1, 0) self.conv_depth = ops.Conv2d(hidden_channels * 2, hidden_channels * 2, 3, 1, 1, groups=hidden_channels * 2) self.conv_point = ops.Conv2d(hidden_channels, out_channels, 1, 1, 0, bias=False) self.norm = None if norm_type == "rms_norm": self.norm = RMSNorm(out_channels, eps=1e-5, elementwise_affine=True, bias=True) def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: if self.residual_connection: residual = hidden_states hidden_states = self.conv_inverted(hidden_states) hidden_states = self.nonlinearity(hidden_states) hidden_states = self.conv_depth(hidden_states) hidden_states, gate = torch.chunk(hidden_states, 2, dim=1) hidden_states = hidden_states * self.nonlinearity(gate) hidden_states = self.conv_point(hidden_states) if self.norm_type == "rms_norm": # move channel to the last dimension so we apply RMSnorm across channel dimension hidden_states = self.norm(hidden_states.movedim(1, -1)).movedim(-1, 1) if self.residual_connection: hidden_states = hidden_states + residual return hidden_states def get_block( block_type: str, in_channels: int, out_channels: int, attention_head_dim: int, norm_type: str, act_fn: str, qkv_mutliscales: tuple = (), ): if block_type == "ResBlock": block = ResBlock(in_channels, out_channels, norm_type, act_fn) elif block_type == "EfficientViTBlock": block = EfficientViTBlock( in_channels, attention_head_dim=attention_head_dim, norm_type=norm_type, qkv_multiscales=qkv_mutliscales ) else: raise ValueError(f"Block with {block_type=} is not supported.") return block class DCDownBlock2d(nn.Module): def __init__(self, in_channels: int, out_channels: int, downsample: bool = False, shortcut: bool = True) -> None: super().__init__() self.downsample = downsample self.factor = 2 self.stride = 1 if downsample else 2 self.group_size = in_channels * self.factor**2 // out_channels self.shortcut = shortcut out_ratio = self.factor**2 if downsample: assert out_channels % out_ratio == 0 out_channels = out_channels // out_ratio self.conv = ops.Conv2d( in_channels, out_channels, kernel_size=3, stride=self.stride, padding=1, ) def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: x = self.conv(hidden_states) if self.downsample: x = F.pixel_unshuffle(x, self.factor) if self.shortcut: y = F.pixel_unshuffle(hidden_states, self.factor) y = y.unflatten(1, (-1, self.group_size)) y = y.mean(dim=2) hidden_states = x + y else: hidden_states = x return hidden_states class DCUpBlock2d(nn.Module): def __init__( self, in_channels: int, out_channels: int, interpolate: bool = False, shortcut: bool = True, interpolation_mode: str = "nearest", ) -> None: super().__init__() self.interpolate = interpolate self.interpolation_mode = interpolation_mode self.shortcut = shortcut self.factor = 2 self.repeats = out_channels * self.factor**2 // in_channels out_ratio = self.factor**2 if not interpolate: out_channels = out_channels * out_ratio self.conv = ops.Conv2d(in_channels, out_channels, 3, 1, 1) def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: if self.interpolate: x = F.interpolate(hidden_states, scale_factor=self.factor, mode=self.interpolation_mode) x = self.conv(x) else: x = self.conv(hidden_states) x = F.pixel_shuffle(x, self.factor) if self.shortcut: y = hidden_states.repeat_interleave(self.repeats, dim=1, output_size=hidden_states.shape[1] * self.repeats) y = F.pixel_shuffle(y, self.factor) hidden_states = x + y else: hidden_states = x return hidden_states class Encoder(nn.Module): def __init__( self, in_channels: int, latent_channels: int, attention_head_dim: int = 32, block_type: str or tuple = "ResBlock", block_out_channels: tuple = (128, 256, 512, 512, 1024, 1024), layers_per_block: tuple = (2, 2, 2, 2, 2, 2), qkv_multiscales: tuple = ((), (), (), (5,), (5,), (5,)), downsample_block_type: str = "pixel_unshuffle", out_shortcut: bool = True, ): super().__init__() num_blocks = len(block_out_channels) if isinstance(block_type, str): block_type = (block_type,) * num_blocks if layers_per_block[0] > 0: self.conv_in = ops.Conv2d( in_channels, block_out_channels[0] if layers_per_block[0] > 0 else block_out_channels[1], kernel_size=3, stride=1, padding=1, ) else: self.conv_in = DCDownBlock2d( in_channels=in_channels, out_channels=block_out_channels[0] if layers_per_block[0] > 0 else block_out_channels[1], downsample=downsample_block_type == "pixel_unshuffle", shortcut=False, ) down_blocks = [] for i, (out_channel, num_layers) in enumerate(zip(block_out_channels, layers_per_block)): down_block_list = [] for _ in range(num_layers): block = get_block( block_type[i], out_channel, out_channel, attention_head_dim=attention_head_dim, norm_type="rms_norm", act_fn="silu", qkv_mutliscales=qkv_multiscales[i], ) down_block_list.append(block) if i < num_blocks - 1 and num_layers > 0: downsample_block = DCDownBlock2d( in_channels=out_channel, out_channels=block_out_channels[i + 1], downsample=downsample_block_type == "pixel_unshuffle", shortcut=True, ) down_block_list.append(downsample_block) down_blocks.append(nn.Sequential(*down_block_list)) self.down_blocks = nn.ModuleList(down_blocks) self.conv_out = ops.Conv2d(block_out_channels[-1], latent_channels, 3, 1, 1) self.out_shortcut = out_shortcut if out_shortcut: self.out_shortcut_average_group_size = block_out_channels[-1] // latent_channels def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: hidden_states = self.conv_in(hidden_states) for down_block in self.down_blocks: hidden_states = down_block(hidden_states) if self.out_shortcut: x = hidden_states.unflatten(1, (-1, self.out_shortcut_average_group_size)) x = x.mean(dim=2) hidden_states = self.conv_out(hidden_states) + x else: hidden_states = self.conv_out(hidden_states) return hidden_states class Decoder(nn.Module): def __init__( self, in_channels: int, latent_channels: int, attention_head_dim: int = 32, block_type: str or tuple = "ResBlock", block_out_channels: tuple = (128, 256, 512, 512, 1024, 1024), layers_per_block: tuple = (2, 2, 2, 2, 2, 2), qkv_multiscales: tuple = ((), (), (), (5,), (5,), (5,)), norm_type: str or tuple = "rms_norm", act_fn: str or tuple = "silu", upsample_block_type: str = "pixel_shuffle", in_shortcut: bool = True, ): super().__init__() num_blocks = len(block_out_channels) if isinstance(block_type, str): block_type = (block_type,) * num_blocks if isinstance(norm_type, str): norm_type = (norm_type,) * num_blocks if isinstance(act_fn, str): act_fn = (act_fn,) * num_blocks self.conv_in = ops.Conv2d(latent_channels, block_out_channels[-1], 3, 1, 1) self.in_shortcut = in_shortcut if in_shortcut: self.in_shortcut_repeats = block_out_channels[-1] // latent_channels up_blocks = [] for i, (out_channel, num_layers) in reversed(list(enumerate(zip(block_out_channels, layers_per_block)))): up_block_list = [] if i < num_blocks - 1 and num_layers > 0: upsample_block = DCUpBlock2d( block_out_channels[i + 1], out_channel, interpolate=upsample_block_type == "interpolate", shortcut=True, ) up_block_list.append(upsample_block) for _ in range(num_layers): block = get_block( block_type[i], out_channel, out_channel, attention_head_dim=attention_head_dim, norm_type=norm_type[i], act_fn=act_fn[i], qkv_mutliscales=qkv_multiscales[i], ) up_block_list.append(block) up_blocks.insert(0, nn.Sequential(*up_block_list)) self.up_blocks = nn.ModuleList(up_blocks) channels = block_out_channels[0] if layers_per_block[0] > 0 else block_out_channels[1] self.norm_out = RMSNorm(channels, 1e-5, elementwise_affine=True, bias=True) self.conv_act = nn.ReLU() self.conv_out = None if layers_per_block[0] > 0: self.conv_out = ops.Conv2d(channels, in_channels, 3, 1, 1) else: self.conv_out = DCUpBlock2d( channels, in_channels, interpolate=upsample_block_type == "interpolate", shortcut=False ) def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: if self.in_shortcut: x = hidden_states.repeat_interleave( self.in_shortcut_repeats, dim=1, output_size=hidden_states.shape[1] * self.in_shortcut_repeats ) hidden_states = self.conv_in(hidden_states) + x else: hidden_states = self.conv_in(hidden_states) for up_block in reversed(self.up_blocks): hidden_states = up_block(hidden_states) hidden_states = self.norm_out(hidden_states.movedim(1, -1)).movedim(-1, 1) hidden_states = self.conv_act(hidden_states) hidden_states = self.conv_out(hidden_states) return hidden_states class AutoencoderDC(nn.Module): def __init__( self, in_channels: int = 2, latent_channels: int = 8, attention_head_dim: int = 32, encoder_block_types: Union[str, Tuple[str]] = ["ResBlock", "ResBlock", "ResBlock", "EfficientViTBlock"], decoder_block_types: Union[str, Tuple[str]] = ["ResBlock", "ResBlock", "ResBlock", "EfficientViTBlock"], encoder_block_out_channels: Tuple[int, ...] = (128, 256, 512, 1024), decoder_block_out_channels: Tuple[int, ...] = (128, 256, 512, 1024), encoder_layers_per_block: Tuple[int] = (2, 2, 3, 3), decoder_layers_per_block: Tuple[int] = (3, 3, 3, 3), encoder_qkv_multiscales: Tuple[Tuple[int, ...], ...] = ((), (), (5,), (5,)), decoder_qkv_multiscales: Tuple[Tuple[int, ...], ...] = ((), (), (5,), (5,)), upsample_block_type: str = "interpolate", downsample_block_type: str = "Conv", decoder_norm_types: Union[str, Tuple[str]] = "rms_norm", decoder_act_fns: Union[str, Tuple[str]] = "silu", scaling_factor: float = 0.41407, ) -> None: super().__init__() self.encoder = Encoder( in_channels=in_channels, latent_channels=latent_channels, attention_head_dim=attention_head_dim, block_type=encoder_block_types, block_out_channels=encoder_block_out_channels, layers_per_block=encoder_layers_per_block, qkv_multiscales=encoder_qkv_multiscales, downsample_block_type=downsample_block_type, ) self.decoder = Decoder( in_channels=in_channels, latent_channels=latent_channels, attention_head_dim=attention_head_dim, block_type=decoder_block_types, block_out_channels=decoder_block_out_channels, layers_per_block=decoder_layers_per_block, qkv_multiscales=decoder_qkv_multiscales, norm_type=decoder_norm_types, act_fn=decoder_act_fns, upsample_block_type=upsample_block_type, ) self.scaling_factor = scaling_factor self.spatial_compression_ratio = 2 ** (len(encoder_block_out_channels) - 1) def encode(self, x: torch.Tensor) -> torch.Tensor: """Internal encoding function.""" encoded = self.encoder(x) return encoded * self.scaling_factor def decode(self, z: torch.Tensor) -> torch.Tensor: # Scale the latents back z = z / self.scaling_factor decoded = self.decoder(z) return decoded def forward(self, x: torch.Tensor) -> torch.Tensor: z = self.encode(x) return self.decode(z)
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/ldm/ace/vae/autoencoder_dc.py", "license": "GNU General Public License v3.0", "lines": 527, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy/ldm/ace/vae/music_dcae_pipeline.py
# Original from: https://github.com/ace-step/ACE-Step/blob/main/music_dcae/music_dcae_pipeline.py import torch from .autoencoder_dc import AutoencoderDC import logging try: import torchaudio except: logging.warning("torchaudio missing, ACE model will be broken") import torchvision.transforms as transforms from .music_vocoder import ADaMoSHiFiGANV1 class MusicDCAE(torch.nn.Module): def __init__(self, source_sample_rate=None, dcae_config={}, vocoder_config={}): super(MusicDCAE, self).__init__() self.dcae = AutoencoderDC(**dcae_config) self.vocoder = ADaMoSHiFiGANV1(**vocoder_config) if source_sample_rate is None: self.source_sample_rate = 48000 else: self.source_sample_rate = source_sample_rate self.transform = transforms.Compose([ transforms.Normalize(0.5, 0.5), ]) self.min_mel_value = -11.0 self.max_mel_value = 3.0 self.audio_chunk_size = int(round((1024 * 512 / 44100 * 48000))) self.mel_chunk_size = 1024 self.time_dimention_multiple = 8 self.latent_chunk_size = self.mel_chunk_size // self.time_dimention_multiple self.scale_factor = 0.1786 self.shift_factor = -1.9091 def forward_mel(self, audios): mels = [] for i in range(len(audios)): image = self.vocoder.mel_transform(audios[i]) mels.append(image) mels = torch.stack(mels) return mels @torch.no_grad() def encode(self, audios, audio_lengths=None, sr=None): if audio_lengths is None: audio_lengths = torch.tensor([audios.shape[2]] * audios.shape[0]) audio_lengths = audio_lengths.to(audios.device) if sr is None: sr = self.source_sample_rate if sr != 44100: audios = torchaudio.functional.resample(audios, sr, 44100) max_audio_len = audios.shape[-1] if max_audio_len % (8 * 512) != 0: audios = torch.nn.functional.pad(audios, (0, 8 * 512 - max_audio_len % (8 * 512))) mels = self.forward_mel(audios) mels = (mels - self.min_mel_value) / (self.max_mel_value - self.min_mel_value) mels = self.transform(mels) latents = [] for mel in mels: latent = self.dcae.encoder(mel.unsqueeze(0)) latents.append(latent) latents = torch.cat(latents, dim=0) latents = (latents - self.shift_factor) * self.scale_factor return latents @torch.no_grad() def decode(self, latents, audio_lengths=None, sr=None): latents = latents / self.scale_factor + self.shift_factor pred_wavs = [] for latent in latents: mels = self.dcae.decoder(latent.unsqueeze(0)) mels = mels * 0.5 + 0.5 mels = mels * (self.max_mel_value - self.min_mel_value) + self.min_mel_value wav = self.vocoder.decode(mels[0]).squeeze(1) if sr is not None: wav = torchaudio.functional.resample(wav, 44100, sr) else: sr = 44100 pred_wavs.append(wav) if audio_lengths is not None: pred_wavs = [wav[:, :length].cpu() for wav, length in zip(pred_wavs, audio_lengths)] return torch.stack(pred_wavs) def forward(self, audios, audio_lengths=None, sr=None): latents, latent_lengths = self.encode(audios=audios, audio_lengths=audio_lengths, sr=sr) sr, pred_wavs = self.decode(latents=latents, audio_lengths=audio_lengths, sr=sr) return sr, pred_wavs, latents, latent_lengths
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/ldm/ace/vae/music_dcae_pipeline.py", "license": "GNU General Public License v3.0", "lines": 80, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy/ldm/ace/vae/music_log_mel.py
# Original from: https://github.com/ace-step/ACE-Step/blob/main/music_dcae/music_log_mel.py import torch import torch.nn as nn from torch import Tensor import logging try: from torchaudio.transforms import MelScale except: logging.warning("torchaudio missing, ACE model will be broken") import comfy.model_management class LinearSpectrogram(nn.Module): def __init__( self, n_fft=2048, win_length=2048, hop_length=512, center=False, mode="pow2_sqrt", ): super().__init__() self.n_fft = n_fft self.win_length = win_length self.hop_length = hop_length self.center = center self.mode = mode self.register_buffer("window", torch.hann_window(win_length)) def forward(self, y: Tensor) -> Tensor: if y.ndim == 3: y = y.squeeze(1) y = torch.nn.functional.pad( y.unsqueeze(1), ( (self.win_length - self.hop_length) // 2, (self.win_length - self.hop_length + 1) // 2, ), mode="reflect", ).squeeze(1) dtype = y.dtype spec = torch.stft( y.float(), self.n_fft, hop_length=self.hop_length, win_length=self.win_length, window=comfy.model_management.cast_to(self.window, dtype=torch.float32, device=y.device), center=self.center, pad_mode="reflect", normalized=False, onesided=True, return_complex=True, ) spec = torch.view_as_real(spec) if self.mode == "pow2_sqrt": spec = torch.sqrt(spec.pow(2).sum(-1) + 1e-6) spec = spec.to(dtype) return spec class LogMelSpectrogram(nn.Module): def __init__( self, sample_rate=44100, n_fft=2048, win_length=2048, hop_length=512, n_mels=128, center=False, f_min=0.0, f_max=None, ): super().__init__() self.sample_rate = sample_rate self.n_fft = n_fft self.win_length = win_length self.hop_length = hop_length self.center = center self.n_mels = n_mels self.f_min = f_min self.f_max = f_max or sample_rate // 2 self.spectrogram = LinearSpectrogram(n_fft, win_length, hop_length, center) self.mel_scale = MelScale( self.n_mels, self.sample_rate, self.f_min, self.f_max, self.n_fft // 2 + 1, "slaney", "slaney", ) def compress(self, x: Tensor) -> Tensor: return torch.log(torch.clamp(x, min=1e-5)) def decompress(self, x: Tensor) -> Tensor: return torch.exp(x) def forward(self, x: Tensor, return_linear: bool = False) -> Tensor: linear = self.spectrogram(x) x = self.mel_scale(linear) x = self.compress(x) # print(x.shape) if return_linear: return x, self.compress(linear) return x
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/ldm/ace/vae/music_log_mel.py", "license": "GNU General Public License v3.0", "lines": 98, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy/text_encoders/ace.py
from comfy import sd1_clip from .spiece_tokenizer import SPieceTokenizer import comfy.text_encoders.t5 import os import re import torch import logging from tokenizers import Tokenizer from .ace_text_cleaners import multilingual_cleaners, japanese_to_romaji SUPPORT_LANGUAGES = { "en": 259, "de": 260, "fr": 262, "es": 284, "it": 285, "pt": 286, "pl": 294, "tr": 295, "ru": 267, "cs": 293, "nl": 297, "ar": 5022, "zh": 5023, "ja": 5412, "hu": 5753, "ko": 6152, "hi": 6680 } structure_pattern = re.compile(r"\[.*?\]") DEFAULT_VOCAB_FILE = os.path.join(os.path.join(os.path.dirname(os.path.realpath(__file__)), "ace_lyrics_tokenizer"), "vocab.json") class VoiceBpeTokenizer: def __init__(self, vocab_file=DEFAULT_VOCAB_FILE): self.tokenizer = None if vocab_file is not None: self.tokenizer = Tokenizer.from_file(vocab_file) def preprocess_text(self, txt, lang): txt = multilingual_cleaners(txt, lang) return txt def encode(self, txt, lang='en'): # lang = lang.split("-")[0] # remove the region # self.check_input_length(txt, lang) txt = self.preprocess_text(txt, lang) lang = "zh-cn" if lang == "zh" else lang txt = f"[{lang}]{txt}" txt = txt.replace(" ", "[SPACE]") return self.tokenizer.encode(txt).ids def get_lang(self, line): if line.startswith("[") and line[3:4] == ']': lang = line[1:3].lower() if lang in SUPPORT_LANGUAGES: return lang, line[4:] return "en", line def __call__(self, string): lines = string.split("\n") lyric_token_idx = [261] for line in lines: line = line.strip() if not line: lyric_token_idx += [2] continue lang, line = self.get_lang(line) if lang not in SUPPORT_LANGUAGES: lang = "en" if "zh" in lang: lang = "zh" if "spa" in lang: lang = "es" try: line_out = japanese_to_romaji(line) if line_out != line: lang = "ja" line = line_out except: pass try: if structure_pattern.match(line): token_idx = self.encode(line, "en") else: token_idx = self.encode(line, lang) lyric_token_idx = lyric_token_idx + token_idx + [2] except Exception as e: logging.warning("tokenize error {} for line {} major_language {}".format(e, line, lang)) return {"input_ids": lyric_token_idx} @staticmethod def from_pretrained(path, **kwargs): return VoiceBpeTokenizer(path, **kwargs) def get_vocab(self): return {} class UMT5BaseModel(sd1_clip.SDClipModel): def __init__(self, device="cpu", layer="last", layer_idx=None, dtype=None, model_options={}): textmodel_json_config = os.path.join(os.path.dirname(os.path.realpath(__file__)), "umt5_config_base.json") super().__init__(device=device, layer=layer, layer_idx=layer_idx, textmodel_json_config=textmodel_json_config, dtype=dtype, special_tokens={"end": 1, "pad": 0}, model_class=comfy.text_encoders.t5.T5, enable_attention_masks=True, zero_out_masked=False, model_options=model_options) class UMT5BaseTokenizer(sd1_clip.SDTokenizer): def __init__(self, embedding_directory=None, tokenizer_data={}): tokenizer = tokenizer_data.get("spiece_model", None) super().__init__(tokenizer, pad_with_end=False, embedding_size=768, embedding_key='umt5base', tokenizer_class=SPieceTokenizer, has_start_token=False, pad_to_max_length=False, max_length=99999999, min_length=1, pad_token=0, tokenizer_data=tokenizer_data) def state_dict(self): return {"spiece_model": self.tokenizer.serialize_model()} class LyricsTokenizer(sd1_clip.SDTokenizer): def __init__(self, embedding_directory=None, tokenizer_data={}): tokenizer = os.path.join(os.path.join(os.path.dirname(os.path.realpath(__file__)), "ace_lyrics_tokenizer"), "vocab.json") super().__init__(tokenizer, pad_with_end=False, embedding_size=1024, embedding_key='lyrics', tokenizer_class=VoiceBpeTokenizer, has_start_token=True, pad_to_max_length=False, max_length=99999999, min_length=1, pad_token=2, has_end_token=False, tokenizer_data=tokenizer_data) class AceT5Tokenizer: def __init__(self, embedding_directory=None, tokenizer_data={}): self.voicebpe = LyricsTokenizer(embedding_directory=embedding_directory, tokenizer_data=tokenizer_data) self.umt5base = UMT5BaseTokenizer(embedding_directory=embedding_directory, tokenizer_data=tokenizer_data) def tokenize_with_weights(self, text:str, return_word_ids=False, **kwargs): out = {} out["lyrics"] = self.voicebpe.tokenize_with_weights(kwargs.get("lyrics", ""), return_word_ids, **kwargs) out["umt5base"] = self.umt5base.tokenize_with_weights(text, return_word_ids, **kwargs) return out def untokenize(self, token_weight_pair): return self.umt5base.untokenize(token_weight_pair) def state_dict(self): return self.umt5base.state_dict() class AceT5Model(torch.nn.Module): def __init__(self, device="cpu", dtype=None, model_options={}, **kwargs): super().__init__() self.umt5base = UMT5BaseModel(device=device, dtype=dtype, model_options=model_options) self.dtypes = set() if dtype is not None: self.dtypes.add(dtype) def set_clip_options(self, options): self.umt5base.set_clip_options(options) def reset_clip_options(self): self.umt5base.reset_clip_options() def encode_token_weights(self, token_weight_pairs): token_weight_pairs_umt5base = token_weight_pairs["umt5base"] token_weight_pairs_lyrics = token_weight_pairs["lyrics"] t5_out, t5_pooled = self.umt5base.encode_token_weights(token_weight_pairs_umt5base) lyrics_embeds = torch.tensor(list(map(lambda a: a[0], token_weight_pairs_lyrics[0]))).unsqueeze(0) return t5_out, None, {"conditioning_lyrics": lyrics_embeds} def load_sd(self, sd): return self.umt5base.load_sd(sd)
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/text_encoders/ace.py", "license": "GNU General Public License v3.0", "lines": 121, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_extras/nodes_ace.py
import torch from typing_extensions import override import comfy.model_management import node_helpers from comfy_api.latest import ComfyExtension, io class TextEncodeAceStepAudio(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="TextEncodeAceStepAudio", category="conditioning", inputs=[ io.Clip.Input("clip"), io.String.Input("tags", multiline=True, dynamic_prompts=True), io.String.Input("lyrics", multiline=True, dynamic_prompts=True), io.Float.Input("lyrics_strength", default=1.0, min=0.0, max=10.0, step=0.01), ], outputs=[io.Conditioning.Output()], ) @classmethod def execute(cls, clip, tags, lyrics, lyrics_strength) -> io.NodeOutput: tokens = clip.tokenize(tags, lyrics=lyrics) conditioning = clip.encode_from_tokens_scheduled(tokens) conditioning = node_helpers.conditioning_set_values(conditioning, {"lyrics_strength": lyrics_strength}) return io.NodeOutput(conditioning) class TextEncodeAceStepAudio15(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="TextEncodeAceStepAudio1.5", category="conditioning", inputs=[ io.Clip.Input("clip"), io.String.Input("tags", multiline=True, dynamic_prompts=True), io.String.Input("lyrics", multiline=True, dynamic_prompts=True), io.Int.Input("seed", default=0, min=0, max=0xffffffffffffffff, control_after_generate=True), io.Int.Input("bpm", default=120, min=10, max=300), io.Float.Input("duration", default=120.0, min=0.0, max=2000.0, step=0.1), io.Combo.Input("timesignature", options=['2', '3', '4', '6']), io.Combo.Input("language", options=["en", "ja", "zh", "es", "de", "fr", "pt", "ru", "it", "nl", "pl", "tr", "vi", "cs", "fa", "id", "ko", "uk", "hu", "ar", "sv", "ro", "el"]), io.Combo.Input("keyscale", options=[f"{root} {quality}" for quality in ["major", "minor"] for root in ["C", "C#", "Db", "D", "D#", "Eb", "E", "F", "F#", "Gb", "G", "G#", "Ab", "A", "A#", "Bb", "B"]]), io.Boolean.Input("generate_audio_codes", default=True, tooltip="Enable the LLM that generates audio codes. This can be slow but will increase the quality of the generated audio. Turn this off if you are giving the model an audio reference.", advanced=True), io.Float.Input("cfg_scale", default=2.0, min=0.0, max=100.0, step=0.1, advanced=True), io.Float.Input("temperature", default=0.85, min=0.0, max=2.0, step=0.01, advanced=True), io.Float.Input("top_p", default=0.9, min=0.0, max=2000.0, step=0.01, advanced=True), io.Int.Input("top_k", default=0, min=0, max=100, advanced=True), io.Float.Input("min_p", default=0.000, min=0.0, max=1.0, step=0.001, advanced=True), ], outputs=[io.Conditioning.Output()], ) @classmethod def execute(cls, clip, tags, lyrics, seed, bpm, duration, timesignature, language, keyscale, generate_audio_codes, cfg_scale, temperature, top_p, top_k, min_p) -> io.NodeOutput: tokens = clip.tokenize(tags, lyrics=lyrics, bpm=bpm, duration=duration, timesignature=int(timesignature), language=language, keyscale=keyscale, seed=seed, generate_audio_codes=generate_audio_codes, cfg_scale=cfg_scale, temperature=temperature, top_p=top_p, top_k=top_k, min_p=min_p) conditioning = clip.encode_from_tokens_scheduled(tokens) return io.NodeOutput(conditioning) class EmptyAceStepLatentAudio(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="EmptyAceStepLatentAudio", display_name="Empty Ace Step 1.0 Latent Audio", category="latent/audio", inputs=[ io.Float.Input("seconds", default=120.0, min=1.0, max=1000.0, step=0.1), io.Int.Input( "batch_size", default=1, min=1, max=4096, tooltip="The number of latent images in the batch." ), ], outputs=[io.Latent.Output()], ) @classmethod def execute(cls, seconds, batch_size) -> io.NodeOutput: length = int(seconds * 44100 / 512 / 8) latent = torch.zeros([batch_size, 8, 16, length], device=comfy.model_management.intermediate_device()) return io.NodeOutput({"samples": latent, "type": "audio"}) class EmptyAceStep15LatentAudio(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="EmptyAceStep1.5LatentAudio", display_name="Empty Ace Step 1.5 Latent Audio", category="latent/audio", inputs=[ io.Float.Input("seconds", default=120.0, min=1.0, max=1000.0, step=0.01), io.Int.Input( "batch_size", default=1, min=1, max=4096, tooltip="The number of latent images in the batch." ), ], outputs=[io.Latent.Output()], ) @classmethod def execute(cls, seconds, batch_size) -> io.NodeOutput: length = round((seconds * 48000 / 1920)) latent = torch.zeros([batch_size, 64, length], device=comfy.model_management.intermediate_device()) return io.NodeOutput({"samples": latent, "type": "audio"}) class ReferenceAudio(io.ComfyNode): @classmethod def define_schema(cls): return io.Schema( node_id="ReferenceTimbreAudio", display_name="Reference Audio", category="advanced/conditioning/audio", is_experimental=True, description="This node sets the reference audio for ace step 1.5", inputs=[ io.Conditioning.Input("conditioning"), io.Latent.Input("latent", optional=True), ], outputs=[ io.Conditioning.Output(), ] ) @classmethod def execute(cls, conditioning, latent=None) -> io.NodeOutput: if latent is not None: conditioning = node_helpers.conditioning_set_values(conditioning, {"reference_audio_timbre_latents": [latent["samples"]]}, append=True) return io.NodeOutput(conditioning) class AceExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[io.ComfyNode]]: return [ TextEncodeAceStepAudio, EmptyAceStepLatentAudio, TextEncodeAceStepAudio15, EmptyAceStep15LatentAudio, ReferenceAudio, ] async def comfy_entrypoint() -> AceExtension: return AceExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_extras/nodes_ace.py", "license": "GNU General Public License v3.0", "lines": 129, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy_api_nodes/nodes_bfl.py
import torch from pydantic import BaseModel from typing_extensions import override from comfy_api.latest import IO, ComfyExtension, Input from comfy_api_nodes.apis.bfl import ( BFLFluxExpandImageRequest, BFLFluxFillImageRequest, BFLFluxKontextProGenerateRequest, BFLFluxProGenerateResponse, BFLFluxProUltraGenerateRequest, BFLFluxStatusResponse, BFLStatus, Flux2ProGenerateRequest, ) from comfy_api_nodes.util import ( ApiEndpoint, download_url_to_image_tensor, get_number_of_images, poll_op, resize_mask_to_image, sync_op, tensor_to_base64_string, validate_aspect_ratio_string, validate_string, ) def convert_mask_to_image(mask: Input.Image): """ Make mask have the expected amount of dims (4) and channels (3) to be recognized as an image. """ mask = mask.unsqueeze(-1) mask = torch.cat([mask] * 3, dim=-1) return mask class FluxProUltraImageNode(IO.ComfyNode): @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id="FluxProUltraImageNode", display_name="Flux 1.1 [pro] Ultra Image", category="api node/image/BFL", description="Generates images using Flux Pro 1.1 Ultra via api based on prompt and resolution.", inputs=[ IO.String.Input( "prompt", multiline=True, default="", tooltip="Prompt for the image generation", ), IO.Boolean.Input( "prompt_upsampling", default=False, tooltip="Whether to perform upsampling on the prompt. " "If active, automatically modifies the prompt for more creative generation, " "but results are nondeterministic (same seed will not produce exactly the same result).", advanced=True, ), IO.Int.Input( "seed", default=0, min=0, max=0xFFFFFFFFFFFFFFFF, control_after_generate=True, tooltip="The random seed used for creating the noise.", ), IO.String.Input( "aspect_ratio", default="16:9", tooltip="Aspect ratio of image; must be between 1:4 and 4:1.", ), IO.Boolean.Input( "raw", default=False, tooltip="When True, generate less processed, more natural-looking images.", ), IO.Image.Input( "image_prompt", optional=True, ), IO.Float.Input( "image_prompt_strength", default=0.1, min=0.0, max=1.0, step=0.01, tooltip="Blend between the prompt and the image prompt.", optional=True, ), ], outputs=[IO.Image.Output()], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.06}""", ), ) @classmethod def validate_inputs(cls, aspect_ratio: str): validate_aspect_ratio_string(aspect_ratio, (1, 4), (4, 1)) return True @classmethod async def execute( cls, prompt: str, aspect_ratio: str, prompt_upsampling: bool = False, raw: bool = False, seed: int = 0, image_prompt: Input.Image | None = None, image_prompt_strength: float = 0.1, ) -> IO.NodeOutput: if image_prompt is None: validate_string(prompt, strip_whitespace=False) initial_response = await sync_op( cls, ApiEndpoint(path="/proxy/bfl/flux-pro-1.1-ultra/generate", method="POST"), response_model=BFLFluxProGenerateResponse, data=BFLFluxProUltraGenerateRequest( prompt=prompt, prompt_upsampling=prompt_upsampling, seed=seed, aspect_ratio=aspect_ratio, raw=raw, image_prompt=(image_prompt if image_prompt is None else tensor_to_base64_string(image_prompt)), image_prompt_strength=(None if image_prompt is None else round(image_prompt_strength, 2)), ), ) response = await poll_op( cls, ApiEndpoint(initial_response.polling_url), response_model=BFLFluxStatusResponse, status_extractor=lambda r: r.status, progress_extractor=lambda r: r.progress, completed_statuses=[BFLStatus.ready], failed_statuses=[ BFLStatus.request_moderated, BFLStatus.content_moderated, BFLStatus.error, BFLStatus.task_not_found, ], queued_statuses=[], ) return IO.NodeOutput(await download_url_to_image_tensor(response.result["sample"])) class FluxKontextProImageNode(IO.ComfyNode): @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id=cls.NODE_ID, display_name=cls.DISPLAY_NAME, category="api node/image/BFL", description="Edits images using Flux.1 Kontext [pro] via api based on prompt and aspect ratio.", inputs=[ IO.String.Input( "prompt", multiline=True, default="", tooltip="Prompt for the image generation - specify what and how to edit.", ), IO.String.Input( "aspect_ratio", default="16:9", tooltip="Aspect ratio of image; must be between 1:4 and 4:1.", ), IO.Float.Input( "guidance", default=3.0, min=0.1, max=99.0, step=0.1, tooltip="Guidance strength for the image generation process", ), IO.Int.Input( "steps", default=50, min=1, max=150, tooltip="Number of steps for the image generation process", ), IO.Int.Input( "seed", default=1234, min=0, max=0xFFFFFFFFFFFFFFFF, control_after_generate=True, tooltip="The random seed used for creating the noise.", ), IO.Boolean.Input( "prompt_upsampling", default=False, tooltip="Whether to perform upsampling on the prompt. If active, automatically modifies the prompt for more creative generation, but results are nondeterministic (same seed will not produce exactly the same result).", advanced=True, ), IO.Image.Input( "input_image", optional=True, ), ], outputs=[IO.Image.Output()], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, ) BFL_PATH = "/proxy/bfl/flux-kontext-pro/generate" NODE_ID = "FluxKontextProImageNode" DISPLAY_NAME = "Flux.1 Kontext [pro] Image" @classmethod async def execute( cls, prompt: str, aspect_ratio: str, guidance: float, steps: int, input_image: Input.Image | None = None, seed=0, prompt_upsampling=False, ) -> IO.NodeOutput: validate_aspect_ratio_string(aspect_ratio, (1, 4), (4, 1)) if input_image is None: validate_string(prompt, strip_whitespace=False) initial_response = await sync_op( cls, ApiEndpoint(path=cls.BFL_PATH, method="POST"), response_model=BFLFluxProGenerateResponse, data=BFLFluxKontextProGenerateRequest( prompt=prompt, prompt_upsampling=prompt_upsampling, guidance=round(guidance, 1), steps=steps, seed=seed, aspect_ratio=aspect_ratio, input_image=(input_image if input_image is None else tensor_to_base64_string(input_image)), ), ) response = await poll_op( cls, ApiEndpoint(initial_response.polling_url), response_model=BFLFluxStatusResponse, status_extractor=lambda r: r.status, progress_extractor=lambda r: r.progress, completed_statuses=[BFLStatus.ready], failed_statuses=[ BFLStatus.request_moderated, BFLStatus.content_moderated, BFLStatus.error, BFLStatus.task_not_found, ], queued_statuses=[], ) return IO.NodeOutput(await download_url_to_image_tensor(response.result["sample"])) class FluxKontextMaxImageNode(FluxKontextProImageNode): DESCRIPTION = "Edits images using Flux.1 Kontext [max] via api based on prompt and aspect ratio." BFL_PATH = "/proxy/bfl/flux-kontext-max/generate" NODE_ID = "FluxKontextMaxImageNode" DISPLAY_NAME = "Flux.1 Kontext [max] Image" class FluxProExpandNode(IO.ComfyNode): @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id="FluxProExpandNode", display_name="Flux.1 Expand Image", category="api node/image/BFL", description="Outpaints image based on prompt.", inputs=[ IO.Image.Input("image"), IO.String.Input( "prompt", multiline=True, default="", tooltip="Prompt for the image generation", ), IO.Boolean.Input( "prompt_upsampling", default=False, tooltip="Whether to perform upsampling on the prompt. " "If active, automatically modifies the prompt for more creative generation, " "but results are nondeterministic (same seed will not produce exactly the same result).", advanced=True, ), IO.Int.Input( "top", default=0, min=0, max=2048, tooltip="Number of pixels to expand at the top of the image", ), IO.Int.Input( "bottom", default=0, min=0, max=2048, tooltip="Number of pixels to expand at the bottom of the image", ), IO.Int.Input( "left", default=0, min=0, max=2048, tooltip="Number of pixels to expand at the left of the image", ), IO.Int.Input( "right", default=0, min=0, max=2048, tooltip="Number of pixels to expand at the right of the image", ), IO.Float.Input( "guidance", default=60, min=1.5, max=100, tooltip="Guidance strength for the image generation process", ), IO.Int.Input( "steps", default=50, min=15, max=50, tooltip="Number of steps for the image generation process", ), IO.Int.Input( "seed", default=0, min=0, max=0xFFFFFFFFFFFFFFFF, control_after_generate=True, tooltip="The random seed used for creating the noise.", ), ], outputs=[IO.Image.Output()], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.05}""", ), ) @classmethod async def execute( cls, image: Input.Image, prompt: str, prompt_upsampling: bool, top: int, bottom: int, left: int, right: int, steps: int, guidance: float, seed=0, ) -> IO.NodeOutput: initial_response = await sync_op( cls, ApiEndpoint(path="/proxy/bfl/flux-pro-1.0-expand/generate", method="POST"), response_model=BFLFluxProGenerateResponse, data=BFLFluxExpandImageRequest( prompt=prompt, prompt_upsampling=prompt_upsampling, top=top, bottom=bottom, left=left, right=right, steps=steps, guidance=guidance, seed=seed, image=tensor_to_base64_string(image), ), ) response = await poll_op( cls, ApiEndpoint(initial_response.polling_url), response_model=BFLFluxStatusResponse, status_extractor=lambda r: r.status, progress_extractor=lambda r: r.progress, completed_statuses=[BFLStatus.ready], failed_statuses=[ BFLStatus.request_moderated, BFLStatus.content_moderated, BFLStatus.error, BFLStatus.task_not_found, ], queued_statuses=[], ) return IO.NodeOutput(await download_url_to_image_tensor(response.result["sample"])) class FluxProFillNode(IO.ComfyNode): @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id="FluxProFillNode", display_name="Flux.1 Fill Image", category="api node/image/BFL", description="Inpaints image based on mask and prompt.", inputs=[ IO.Image.Input("image"), IO.Mask.Input("mask"), IO.String.Input( "prompt", multiline=True, default="", tooltip="Prompt for the image generation", ), IO.Boolean.Input( "prompt_upsampling", default=False, tooltip="Whether to perform upsampling on the prompt. " "If active, automatically modifies the prompt for more creative generation, " "but results are nondeterministic (same seed will not produce exactly the same result).", advanced=True, ), IO.Float.Input( "guidance", default=60, min=1.5, max=100, tooltip="Guidance strength for the image generation process", ), IO.Int.Input( "steps", default=50, min=15, max=50, tooltip="Number of steps for the image generation process", ), IO.Int.Input( "seed", default=0, min=0, max=0xFFFFFFFFFFFFFFFF, control_after_generate=True, tooltip="The random seed used for creating the noise.", ), ], outputs=[IO.Image.Output()], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.05}""", ), ) @classmethod async def execute( cls, image: Input.Image, mask: Input.Image, prompt: str, prompt_upsampling: bool, steps: int, guidance: float, seed=0, ) -> IO.NodeOutput: # prepare mask mask = resize_mask_to_image(mask, image) mask = tensor_to_base64_string(convert_mask_to_image(mask)) initial_response = await sync_op( cls, ApiEndpoint(path="/proxy/bfl/flux-pro-1.0-fill/generate", method="POST"), response_model=BFLFluxProGenerateResponse, data=BFLFluxFillImageRequest( prompt=prompt, prompt_upsampling=prompt_upsampling, steps=steps, guidance=guidance, seed=seed, image=tensor_to_base64_string(image[:, :, :, :3]), # make sure image will have alpha channel removed mask=mask, ), ) response = await poll_op( cls, ApiEndpoint(initial_response.polling_url), response_model=BFLFluxStatusResponse, status_extractor=lambda r: r.status, progress_extractor=lambda r: r.progress, completed_statuses=[BFLStatus.ready], failed_statuses=[ BFLStatus.request_moderated, BFLStatus.content_moderated, BFLStatus.error, BFLStatus.task_not_found, ], queued_statuses=[], ) return IO.NodeOutput(await download_url_to_image_tensor(response.result["sample"])) class Flux2ProImageNode(IO.ComfyNode): NODE_ID = "Flux2ProImageNode" DISPLAY_NAME = "Flux.2 [pro] Image" API_ENDPOINT = "/proxy/bfl/flux-2-pro/generate" PRICE_BADGE_EXPR = """ ( $MP := 1024 * 1024; $outMP := $max([1, $floor(((widgets.width * widgets.height) + $MP - 1) / $MP)]); $outputCost := 0.03 + 0.015 * ($outMP - 1); inputs.images.connected ? { "type":"range_usd", "min_usd": $outputCost + 0.015, "max_usd": $outputCost + 0.12, "format": { "approximate": true } } : {"type":"usd","usd": $outputCost} ) """ @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id=cls.NODE_ID, display_name=cls.DISPLAY_NAME, category="api node/image/BFL", description="Generates images synchronously based on prompt and resolution.", inputs=[ IO.String.Input( "prompt", multiline=True, default="", tooltip="Prompt for the image generation or edit", ), IO.Int.Input( "width", default=1024, min=256, max=2048, step=32, ), IO.Int.Input( "height", default=768, min=256, max=2048, step=32, ), IO.Int.Input( "seed", default=0, min=0, max=0xFFFFFFFFFFFFFFFF, control_after_generate=True, tooltip="The random seed used for creating the noise.", ), IO.Boolean.Input( "prompt_upsampling", default=True, tooltip="Whether to perform upsampling on the prompt. " "If active, automatically modifies the prompt for more creative generation.", advanced=True, ), IO.Image.Input("images", optional=True, tooltip="Up to 9 images to be used as references."), ], outputs=[IO.Image.Output()], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["width", "height"], inputs=["images"]), expr=cls.PRICE_BADGE_EXPR, ), ) @classmethod async def execute( cls, prompt: str, width: int, height: int, seed: int, prompt_upsampling: bool, images: Input.Image | None = None, ) -> IO.NodeOutput: reference_images = {} if images is not None: if get_number_of_images(images) > 9: raise ValueError("The current maximum number of supported images is 9.") for image_index in range(images.shape[0]): key_name = f"input_image_{image_index + 1}" if image_index else "input_image" reference_images[key_name] = tensor_to_base64_string(images[image_index], total_pixels=2048 * 2048) initial_response = await sync_op( cls, ApiEndpoint(path=cls.API_ENDPOINT, method="POST"), response_model=BFLFluxProGenerateResponse, data=Flux2ProGenerateRequest( prompt=prompt, width=width, height=height, seed=seed, prompt_upsampling=prompt_upsampling, **reference_images, ), ) def price_extractor(_r: BaseModel) -> float | None: return None if initial_response.cost is None else initial_response.cost / 100 response = await poll_op( cls, ApiEndpoint(initial_response.polling_url), response_model=BFLFluxStatusResponse, status_extractor=lambda r: r.status, progress_extractor=lambda r: r.progress, price_extractor=price_extractor, completed_statuses=[BFLStatus.ready], failed_statuses=[ BFLStatus.request_moderated, BFLStatus.content_moderated, BFLStatus.error, BFLStatus.task_not_found, ], queued_statuses=[], ) return IO.NodeOutput(await download_url_to_image_tensor(response.result["sample"])) class Flux2MaxImageNode(Flux2ProImageNode): NODE_ID = "Flux2MaxImageNode" DISPLAY_NAME = "Flux.2 [max] Image" API_ENDPOINT = "/proxy/bfl/flux-2-max/generate" PRICE_BADGE_EXPR = """ ( $MP := 1024 * 1024; $outMP := $max([1, $floor(((widgets.width * widgets.height) + $MP - 1) / $MP)]); $outputCost := 0.07 + 0.03 * ($outMP - 1); inputs.images.connected ? { "type":"range_usd", "min_usd": $outputCost + 0.03, "max_usd": $outputCost + 0.24, "format": { "approximate": true } } : {"type":"usd","usd": $outputCost} ) """ class BFLExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[IO.ComfyNode]]: return [ FluxProUltraImageNode, FluxKontextProImageNode, FluxKontextMaxImageNode, FluxProExpandNode, FluxProFillNode, Flux2ProImageNode, Flux2MaxImageNode, ] async def comfy_entrypoint() -> BFLExtension: return BFLExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api_nodes/nodes_bfl.py", "license": "GNU General Public License v3.0", "lines": 653, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_api_nodes/nodes_ideogram.py
from io import BytesIO from typing_extensions import override from comfy_api.latest import IO, ComfyExtension from PIL import Image import numpy as np import torch from comfy_api_nodes.apis.ideogram import ( IdeogramGenerateRequest, IdeogramGenerateResponse, ImageRequest, IdeogramV3Request, IdeogramV3EditRequest, ) from comfy_api_nodes.util import ( ApiEndpoint, bytesio_to_image_tensor, download_url_as_bytesio, resize_mask_to_image, sync_op, ) V1_V1_RES_MAP = { "Auto":"AUTO", "512 x 1536":"RESOLUTION_512_1536", "576 x 1408":"RESOLUTION_576_1408", "576 x 1472":"RESOLUTION_576_1472", "576 x 1536":"RESOLUTION_576_1536", "640 x 1024":"RESOLUTION_640_1024", "640 x 1344":"RESOLUTION_640_1344", "640 x 1408":"RESOLUTION_640_1408", "640 x 1472":"RESOLUTION_640_1472", "640 x 1536":"RESOLUTION_640_1536", "704 x 1152":"RESOLUTION_704_1152", "704 x 1216":"RESOLUTION_704_1216", "704 x 1280":"RESOLUTION_704_1280", "704 x 1344":"RESOLUTION_704_1344", "704 x 1408":"RESOLUTION_704_1408", "704 x 1472":"RESOLUTION_704_1472", "720 x 1280":"RESOLUTION_720_1280", "736 x 1312":"RESOLUTION_736_1312", "768 x 1024":"RESOLUTION_768_1024", "768 x 1088":"RESOLUTION_768_1088", "768 x 1152":"RESOLUTION_768_1152", "768 x 1216":"RESOLUTION_768_1216", "768 x 1232":"RESOLUTION_768_1232", "768 x 1280":"RESOLUTION_768_1280", "768 x 1344":"RESOLUTION_768_1344", "832 x 960":"RESOLUTION_832_960", "832 x 1024":"RESOLUTION_832_1024", "832 x 1088":"RESOLUTION_832_1088", "832 x 1152":"RESOLUTION_832_1152", "832 x 1216":"RESOLUTION_832_1216", "832 x 1248":"RESOLUTION_832_1248", "864 x 1152":"RESOLUTION_864_1152", "896 x 960":"RESOLUTION_896_960", "896 x 1024":"RESOLUTION_896_1024", "896 x 1088":"RESOLUTION_896_1088", "896 x 1120":"RESOLUTION_896_1120", "896 x 1152":"RESOLUTION_896_1152", "960 x 832":"RESOLUTION_960_832", "960 x 896":"RESOLUTION_960_896", "960 x 1024":"RESOLUTION_960_1024", "960 x 1088":"RESOLUTION_960_1088", "1024 x 640":"RESOLUTION_1024_640", "1024 x 768":"RESOLUTION_1024_768", "1024 x 832":"RESOLUTION_1024_832", "1024 x 896":"RESOLUTION_1024_896", "1024 x 960":"RESOLUTION_1024_960", "1024 x 1024":"RESOLUTION_1024_1024", "1088 x 768":"RESOLUTION_1088_768", "1088 x 832":"RESOLUTION_1088_832", "1088 x 896":"RESOLUTION_1088_896", "1088 x 960":"RESOLUTION_1088_960", "1120 x 896":"RESOLUTION_1120_896", "1152 x 704":"RESOLUTION_1152_704", "1152 x 768":"RESOLUTION_1152_768", "1152 x 832":"RESOLUTION_1152_832", "1152 x 864":"RESOLUTION_1152_864", "1152 x 896":"RESOLUTION_1152_896", "1216 x 704":"RESOLUTION_1216_704", "1216 x 768":"RESOLUTION_1216_768", "1216 x 832":"RESOLUTION_1216_832", "1232 x 768":"RESOLUTION_1232_768", "1248 x 832":"RESOLUTION_1248_832", "1280 x 704":"RESOLUTION_1280_704", "1280 x 720":"RESOLUTION_1280_720", "1280 x 768":"RESOLUTION_1280_768", "1280 x 800":"RESOLUTION_1280_800", "1312 x 736":"RESOLUTION_1312_736", "1344 x 640":"RESOLUTION_1344_640", "1344 x 704":"RESOLUTION_1344_704", "1344 x 768":"RESOLUTION_1344_768", "1408 x 576":"RESOLUTION_1408_576", "1408 x 640":"RESOLUTION_1408_640", "1408 x 704":"RESOLUTION_1408_704", "1472 x 576":"RESOLUTION_1472_576", "1472 x 640":"RESOLUTION_1472_640", "1472 x 704":"RESOLUTION_1472_704", "1536 x 512":"RESOLUTION_1536_512", "1536 x 576":"RESOLUTION_1536_576", "1536 x 640":"RESOLUTION_1536_640", } V1_V2_RATIO_MAP = { "1:1":"ASPECT_1_1", "4:3":"ASPECT_4_3", "3:4":"ASPECT_3_4", "16:9":"ASPECT_16_9", "9:16":"ASPECT_9_16", "2:1":"ASPECT_2_1", "1:2":"ASPECT_1_2", "3:2":"ASPECT_3_2", "2:3":"ASPECT_2_3", "4:5":"ASPECT_4_5", "5:4":"ASPECT_5_4", } V3_RATIO_MAP = { "1:3":"1x3", "3:1":"3x1", "1:2":"1x2", "2:1":"2x1", "9:16":"9x16", "16:9":"16x9", "10:16":"10x16", "16:10":"16x10", "2:3":"2x3", "3:2":"3x2", "3:4":"3x4", "4:3":"4x3", "4:5":"4x5", "5:4":"5x4", "1:1":"1x1", } V3_RESOLUTIONS= [ "Auto", "512x1536", "576x1408", "576x1472", "576x1536", "640x1344", "640x1408", "640x1472", "640x1536", "704x1152", "704x1216", "704x1280", "704x1344", "704x1408", "704x1472", "736x1312", "768x1088", "768x1216", "768x1280", "768x1344", "800x1280", "832x960", "832x1024", "832x1088", "832x1152", "832x1216", "832x1248", "864x1152", "896x960", "896x1024", "896x1088", "896x1120", "896x1152", "960x832", "960x896", "960x1024", "960x1088", "1024x832", "1024x896", "1024x960", "1024x1024", "1088x768", "1088x832", "1088x896", "1088x960", "1120x896", "1152x704", "1152x832", "1152x864", "1152x896", "1216x704", "1216x768", "1216x832", "1248x832", "1280x704", "1280x768", "1280x800", "1312x736", "1344x640", "1344x704", "1344x768", "1408x576", "1408x640", "1408x704", "1472x576", "1472x640", "1472x704", "1536x512", "1536x576", "1536x640" ] async def download_and_process_images(image_urls): """Helper function to download and process multiple images from URLs""" # Initialize list to store image tensors image_tensors = [] for image_url in image_urls: # Using functions from apinode_utils.py to handle downloading and processing image_bytesio = await download_url_as_bytesio(image_url) # Download image content to BytesIO img_tensor = bytesio_to_image_tensor(image_bytesio, mode="RGB") # Convert to torch.Tensor with RGB mode image_tensors.append(img_tensor) # Stack tensors to match (N, width, height, channels) if image_tensors: stacked_tensors = torch.cat(image_tensors, dim=0) else: raise Exception("No valid images were processed") return stacked_tensors class IdeogramV1(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="IdeogramV1", display_name="Ideogram V1", category="api node/image/Ideogram", description="Generates images using the Ideogram V1 model.", inputs=[ IO.String.Input( "prompt", multiline=True, default="", tooltip="Prompt for the image generation", ), IO.Boolean.Input( "turbo", default=False, tooltip="Whether to use turbo mode (faster generation, potentially lower quality)", ), IO.Combo.Input( "aspect_ratio", options=list(V1_V2_RATIO_MAP.keys()), default="1:1", tooltip="The aspect ratio for image generation.", optional=True, ), IO.Combo.Input( "magic_prompt_option", options=["AUTO", "ON", "OFF"], default="AUTO", tooltip="Determine if MagicPrompt should be used in generation", optional=True, advanced=True, ), IO.Int.Input( "seed", default=0, min=0, max=2147483647, step=1, control_after_generate=True, display_mode=IO.NumberDisplay.number, optional=True, ), IO.String.Input( "negative_prompt", multiline=True, default="", tooltip="Description of what to exclude from the image", optional=True, ), IO.Int.Input( "num_images", default=1, min=1, max=8, step=1, display_mode=IO.NumberDisplay.number, optional=True, ), ], outputs=[ IO.Image.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["num_images", "turbo"]), expr=""" ( $n := widgets.num_images; $base := (widgets.turbo = true) ? 0.0286 : 0.0858; {"type":"usd","usd": $round($base * $n, 2)} ) """, ), ) @classmethod async def execute( cls, prompt, turbo=False, aspect_ratio="1:1", magic_prompt_option="AUTO", seed=0, negative_prompt="", num_images=1, ): # Determine the model based on turbo setting aspect_ratio = V1_V2_RATIO_MAP.get(aspect_ratio, None) model = "V_1_TURBO" if turbo else "V_1" response = await sync_op( cls, ApiEndpoint(path="/proxy/ideogram/generate", method="POST"), response_model=IdeogramGenerateResponse, data=IdeogramGenerateRequest( image_request=ImageRequest( prompt=prompt, model=model, num_images=num_images, seed=seed, aspect_ratio=aspect_ratio if aspect_ratio != "ASPECT_1_1" else None, magic_prompt_option=(magic_prompt_option if magic_prompt_option != "AUTO" else None), negative_prompt=negative_prompt if negative_prompt else None, ) ), max_retries=1, ) if not response.data or len(response.data) == 0: raise Exception("No images were generated in the response") image_urls = [image_data.url for image_data in response.data if image_data.url] if not image_urls: raise Exception("No image URLs were generated in the response") return IO.NodeOutput(await download_and_process_images(image_urls)) class IdeogramV2(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="IdeogramV2", display_name="Ideogram V2", category="api node/image/Ideogram", description="Generates images using the Ideogram V2 model.", inputs=[ IO.String.Input( "prompt", multiline=True, default="", tooltip="Prompt for the image generation", ), IO.Boolean.Input( "turbo", default=False, tooltip="Whether to use turbo mode (faster generation, potentially lower quality)", ), IO.Combo.Input( "aspect_ratio", options=list(V1_V2_RATIO_MAP.keys()), default="1:1", tooltip="The aspect ratio for image generation. Ignored if resolution is not set to AUTO.", optional=True, ), IO.Combo.Input( "resolution", options=list(V1_V1_RES_MAP.keys()), default="Auto", tooltip="The resolution for image generation. " "If not set to AUTO, this overrides the aspect_ratio setting.", optional=True, ), IO.Combo.Input( "magic_prompt_option", options=["AUTO", "ON", "OFF"], default="AUTO", tooltip="Determine if MagicPrompt should be used in generation", optional=True, advanced=True, ), IO.Int.Input( "seed", default=0, min=0, max=2147483647, step=1, control_after_generate=True, display_mode=IO.NumberDisplay.number, optional=True, ), IO.Combo.Input( "style_type", options=["AUTO", "GENERAL", "REALISTIC", "DESIGN", "RENDER_3D", "ANIME"], default="NONE", tooltip="Style type for generation (V2 only)", optional=True, advanced=True, ), IO.String.Input( "negative_prompt", multiline=True, default="", tooltip="Description of what to exclude from the image", optional=True, ), IO.Int.Input( "num_images", default=1, min=1, max=8, step=1, display_mode=IO.NumberDisplay.number, optional=True, ), #"color_palette": ( # IO.STRING, # { # "multiline": False, # "default": "", # "tooltip": "Color palette preset name or hex colors with weights", # }, #), ], outputs=[ IO.Image.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["num_images", "turbo"]), expr=""" ( $n := widgets.num_images; $base := (widgets.turbo = true) ? 0.0715 : 0.1144; {"type":"usd","usd": $round($base * $n, 2)} ) """, ), ) @classmethod async def execute( cls, prompt, turbo=False, aspect_ratio="1:1", resolution="Auto", magic_prompt_option="AUTO", seed=0, style_type="NONE", negative_prompt="", num_images=1, color_palette="", ): aspect_ratio = V1_V2_RATIO_MAP.get(aspect_ratio, None) resolution = V1_V1_RES_MAP.get(resolution, None) # Determine the model based on turbo setting model = "V_2_TURBO" if turbo else "V_2" # Handle resolution vs aspect_ratio logic # If resolution is not AUTO, it overrides aspect_ratio final_resolution = None final_aspect_ratio = None if resolution != "AUTO": final_resolution = resolution else: final_aspect_ratio = aspect_ratio if aspect_ratio != "ASPECT_1_1" else None response = await sync_op( cls, endpoint=ApiEndpoint(path="/proxy/ideogram/generate", method="POST"), response_model=IdeogramGenerateResponse, data=IdeogramGenerateRequest( image_request=ImageRequest( prompt=prompt, model=model, num_images=num_images, seed=seed, aspect_ratio=final_aspect_ratio, resolution=final_resolution, magic_prompt_option=(magic_prompt_option if magic_prompt_option != "AUTO" else None), style_type=style_type if style_type != "NONE" else None, negative_prompt=negative_prompt if negative_prompt else None, color_palette=color_palette if color_palette else None, ) ), max_retries=1, ) if not response.data or len(response.data) == 0: raise Exception("No images were generated in the response") image_urls = [image_data.url for image_data in response.data if image_data.url] if not image_urls: raise Exception("No image URLs were generated in the response") return IO.NodeOutput(await download_and_process_images(image_urls)) class IdeogramV3(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="IdeogramV3", display_name="Ideogram V3", category="api node/image/Ideogram", description="Generates images using the Ideogram V3 model. " "Supports both regular image generation from text prompts and image editing with mask.", inputs=[ IO.String.Input( "prompt", multiline=True, default="", tooltip="Prompt for the image generation or editing", ), IO.Image.Input( "image", tooltip="Optional reference image for image editing.", optional=True, ), IO.Mask.Input( "mask", tooltip="Optional mask for inpainting (white areas will be replaced)", optional=True, ), IO.Combo.Input( "aspect_ratio", options=list(V3_RATIO_MAP.keys()), default="1:1", tooltip="The aspect ratio for image generation. Ignored if resolution is not set to Auto.", optional=True, ), IO.Combo.Input( "resolution", options=V3_RESOLUTIONS, default="Auto", tooltip="The resolution for image generation. " "If not set to Auto, this overrides the aspect_ratio setting.", optional=True, ), IO.Combo.Input( "magic_prompt_option", options=["AUTO", "ON", "OFF"], default="AUTO", tooltip="Determine if MagicPrompt should be used in generation", optional=True, advanced=True, ), IO.Int.Input( "seed", default=0, min=0, max=2147483647, step=1, control_after_generate=True, display_mode=IO.NumberDisplay.number, optional=True, ), IO.Int.Input( "num_images", default=1, min=1, max=8, step=1, display_mode=IO.NumberDisplay.number, optional=True, ), IO.Combo.Input( "rendering_speed", options=["DEFAULT", "TURBO", "QUALITY"], default="DEFAULT", tooltip="Controls the trade-off between generation speed and quality", optional=True, advanced=True, ), IO.Image.Input( "character_image", tooltip="Image to use as character reference.", optional=True, ), IO.Mask.Input( "character_mask", tooltip="Optional mask for character reference image.", optional=True, ), ], outputs=[ IO.Image.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["rendering_speed", "num_images"], inputs=["character_image"]), expr=""" ( $n := widgets.num_images; $speed := widgets.rendering_speed; $hasChar := inputs.character_image.connected; $base := $contains($speed,"quality") ? ($hasChar ? 0.286 : 0.1287) : $contains($speed,"default") ? ($hasChar ? 0.2145 : 0.0858) : $contains($speed,"turbo") ? ($hasChar ? 0.143 : 0.0429) : 0.0858; {"type":"usd","usd": $round($base * $n, 2)} ) """, ), ) @classmethod async def execute( cls, prompt, image=None, mask=None, resolution="Auto", aspect_ratio="1:1", magic_prompt_option="AUTO", seed=0, num_images=1, rendering_speed="DEFAULT", character_image=None, character_mask=None, ): if rendering_speed == "BALANCED": # for backward compatibility rendering_speed = "DEFAULT" character_img_binary = None character_mask_binary = None if character_image is not None: input_tensor = character_image.squeeze().cpu() if character_mask is not None: character_mask = resize_mask_to_image(character_mask, character_image, allow_gradient=False) character_mask = 1.0 - character_mask if character_mask.shape[1:] != character_image.shape[1:-1]: raise Exception("Character mask and image must be the same size") mask_np = (character_mask.squeeze().cpu().numpy() * 255).astype(np.uint8) mask_img = Image.fromarray(mask_np) mask_byte_arr = BytesIO() mask_img.save(mask_byte_arr, format="PNG") mask_byte_arr.seek(0) character_mask_binary = mask_byte_arr character_mask_binary.name = "mask.png" img_np = (input_tensor.numpy() * 255).astype(np.uint8) img = Image.fromarray(img_np) img_byte_arr = BytesIO() img.save(img_byte_arr, format="PNG") img_byte_arr.seek(0) character_img_binary = img_byte_arr character_img_binary.name = "image.png" elif character_mask is not None: raise Exception("Character mask requires character image to be present") # Check if both image and mask are provided for editing mode if image is not None and mask is not None: # Process image and mask input_tensor = image.squeeze().cpu() # Resize mask to match image dimension mask = resize_mask_to_image(mask, image, allow_gradient=False) # Invert mask, as Ideogram API will edit black areas instead of white areas (opposite of convention). mask = 1.0 - mask # Validate mask dimensions match image if mask.shape[1:] != image.shape[1:-1]: raise Exception("Mask and Image must be the same size") # Process image img_np = (input_tensor.numpy() * 255).astype(np.uint8) img = Image.fromarray(img_np) img_byte_arr = BytesIO() img.save(img_byte_arr, format="PNG") img_byte_arr.seek(0) img_binary = img_byte_arr img_binary.name = "image.png" # Process mask - white areas will be replaced mask_np = (mask.squeeze().cpu().numpy() * 255).astype(np.uint8) mask_img = Image.fromarray(mask_np) mask_byte_arr = BytesIO() mask_img.save(mask_byte_arr, format="PNG") mask_byte_arr.seek(0) mask_binary = mask_byte_arr mask_binary.name = "mask.png" # Create edit request edit_request = IdeogramV3EditRequest( prompt=prompt, rendering_speed=rendering_speed, ) # Add optional parameters if magic_prompt_option != "AUTO": edit_request.magic_prompt = magic_prompt_option if seed != 0: edit_request.seed = seed if num_images > 1: edit_request.num_images = num_images files = { "image": img_binary, "mask": mask_binary, } if character_img_binary: files["character_reference_images"] = character_img_binary if character_mask_binary: files["character_mask_binary"] = character_mask_binary response = await sync_op( cls, ApiEndpoint(path="/proxy/ideogram/ideogram-v3/edit", method="POST"), response_model=IdeogramGenerateResponse, data=edit_request, files=files, content_type="multipart/form-data", max_retries=1, ) elif image is not None or mask is not None: # If only one of image or mask is provided, raise an error raise Exception("Ideogram V3 image editing requires both an image AND a mask") else: # Create generation request gen_request = IdeogramV3Request( prompt=prompt, rendering_speed=rendering_speed, ) # Handle resolution vs aspect ratio if resolution != "Auto": gen_request.resolution = resolution elif aspect_ratio != "1:1": v3_aspect = V3_RATIO_MAP.get(aspect_ratio) if v3_aspect: gen_request.aspect_ratio = v3_aspect # Add optional parameters if magic_prompt_option != "AUTO": gen_request.magic_prompt = magic_prompt_option if seed != 0: gen_request.seed = seed if num_images > 1: gen_request.num_images = num_images files = {} if character_img_binary: files["character_reference_images"] = character_img_binary if character_mask_binary: files["character_mask_binary"] = character_mask_binary if files: gen_request.style_type = "AUTO" response = await sync_op( cls, endpoint=ApiEndpoint(path="/proxy/ideogram/ideogram-v3/generate", method="POST"), response_model=IdeogramGenerateResponse, data=gen_request, files=files if files else None, content_type="multipart/form-data", max_retries=1, ) if not response.data or len(response.data) == 0: raise Exception("No images were generated in the response") image_urls = [image_data.url for image_data in response.data if image_data.url] if not image_urls: raise Exception("No image URLs were generated in the response") return IO.NodeOutput(await download_and_process_images(image_urls)) class IdeogramExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[IO.ComfyNode]]: return [ IdeogramV1, IdeogramV2, IdeogramV3, ] async def comfy_entrypoint() -> IdeogramExtension: return IdeogramExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api_nodes/nodes_ideogram.py", "license": "GNU General Public License v3.0", "lines": 761, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_api_nodes/nodes_luma.py
from typing import Optional import torch from typing_extensions import override from comfy_api.latest import IO, ComfyExtension from comfy_api_nodes.apis.luma import ( LumaAspectRatio, LumaCharacterRef, LumaConceptChain, LumaGeneration, LumaGenerationRequest, LumaImageGenerationRequest, LumaImageIdentity, LumaImageModel, LumaImageReference, LumaIO, LumaKeyframes, LumaModifyImageRef, LumaReference, LumaReferenceChain, LumaVideoModel, LumaVideoModelOutputDuration, LumaVideoOutputResolution, get_luma_concepts, ) from comfy_api_nodes.util import ( ApiEndpoint, download_url_to_image_tensor, download_url_to_video_output, poll_op, sync_op, upload_images_to_comfyapi, validate_string, ) LUMA_T2V_AVERAGE_DURATION = 105 LUMA_I2V_AVERAGE_DURATION = 100 class LumaReferenceNode(IO.ComfyNode): @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id="LumaReferenceNode", display_name="Luma Reference", category="api node/image/Luma", description="Holds an image and weight for use with Luma Generate Image node.", inputs=[ IO.Image.Input( "image", tooltip="Image to use as reference.", ), IO.Float.Input( "weight", default=1.0, min=0.0, max=1.0, step=0.01, tooltip="Weight of image reference.", ), IO.Custom(LumaIO.LUMA_REF).Input( "luma_ref", optional=True, ), ], outputs=[IO.Custom(LumaIO.LUMA_REF).Output(display_name="luma_ref")], ) @classmethod def execute(cls, image: torch.Tensor, weight: float, luma_ref: LumaReferenceChain = None) -> IO.NodeOutput: if luma_ref is not None: luma_ref = luma_ref.clone() else: luma_ref = LumaReferenceChain() luma_ref.add(LumaReference(image=image, weight=round(weight, 2))) return IO.NodeOutput(luma_ref) class LumaConceptsNode(IO.ComfyNode): @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id="LumaConceptsNode", display_name="Luma Concepts", category="api node/video/Luma", description="Camera Concepts for use with Luma Text to Video and Luma Image to Video nodes.", inputs=[ IO.Combo.Input( "concept1", options=get_luma_concepts(include_none=True), ), IO.Combo.Input( "concept2", options=get_luma_concepts(include_none=True), ), IO.Combo.Input( "concept3", options=get_luma_concepts(include_none=True), ), IO.Combo.Input( "concept4", options=get_luma_concepts(include_none=True), ), IO.Custom(LumaIO.LUMA_CONCEPTS).Input( "luma_concepts", tooltip="Optional Camera Concepts to add to the ones chosen here.", optional=True, ), ], outputs=[IO.Custom(LumaIO.LUMA_CONCEPTS).Output(display_name="luma_concepts")], ) @classmethod def execute( cls, concept1: str, concept2: str, concept3: str, concept4: str, luma_concepts: LumaConceptChain = None, ) -> IO.NodeOutput: chain = LumaConceptChain(str_list=[concept1, concept2, concept3, concept4]) if luma_concepts is not None: chain = luma_concepts.clone_and_merge(chain) return IO.NodeOutput(chain) class LumaImageGenerationNode(IO.ComfyNode): @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id="LumaImageNode", display_name="Luma Text to Image", category="api node/image/Luma", description="Generates images synchronously based on prompt and aspect ratio.", inputs=[ IO.String.Input( "prompt", multiline=True, default="", tooltip="Prompt for the image generation", ), IO.Combo.Input( "model", options=LumaImageModel, ), IO.Combo.Input( "aspect_ratio", options=LumaAspectRatio, default=LumaAspectRatio.ratio_16_9, ), IO.Int.Input( "seed", default=0, min=0, max=0xFFFFFFFFFFFFFFFF, control_after_generate=True, tooltip="Seed to determine if node should re-run; actual results are nondeterministic regardless of seed.", ), IO.Float.Input( "style_image_weight", default=1.0, min=0.0, max=1.0, step=0.01, tooltip="Weight of style image. Ignored if no style_image provided.", ), IO.Custom(LumaIO.LUMA_REF).Input( "image_luma_ref", tooltip="Luma Reference node connection to influence generation with input images; up to 4 images can be considered.", optional=True, ), IO.Image.Input( "style_image", tooltip="Style reference image; only 1 image will be used.", optional=True, ), IO.Image.Input( "character_image", tooltip="Character reference images; can be a batch of multiple, up to 4 images can be considered.", optional=True, ), ], outputs=[IO.Image.Output()], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["model"]), expr=""" ( $m := widgets.model; $contains($m,"photon-flash-1") ? {"type":"usd","usd":0.0027} : $contains($m,"photon-1") ? {"type":"usd","usd":0.0104} : {"type":"usd","usd":0.0246} ) """, ), ) @classmethod async def execute( cls, prompt: str, model: str, aspect_ratio: str, seed, style_image_weight: float, image_luma_ref: Optional[LumaReferenceChain] = None, style_image: Optional[torch.Tensor] = None, character_image: Optional[torch.Tensor] = None, ) -> IO.NodeOutput: validate_string(prompt, strip_whitespace=True, min_length=3) # handle image_luma_ref api_image_ref = None if image_luma_ref is not None: api_image_ref = await cls._convert_luma_refs(image_luma_ref, max_refs=4) # handle style_luma_ref api_style_ref = None if style_image is not None: api_style_ref = await cls._convert_style_image(style_image, weight=style_image_weight) # handle character_ref images character_ref = None if character_image is not None: download_urls = await upload_images_to_comfyapi(cls, character_image, max_images=4) character_ref = LumaCharacterRef(identity0=LumaImageIdentity(images=download_urls)) response_api = await sync_op( cls, ApiEndpoint(path="/proxy/luma/generations/image", method="POST"), response_model=LumaGeneration, data=LumaImageGenerationRequest( prompt=prompt, model=model, aspect_ratio=aspect_ratio, image_ref=api_image_ref, style_ref=api_style_ref, character_ref=character_ref, ), ) response_poll = await poll_op( cls, ApiEndpoint(path=f"/proxy/luma/generations/{response_api.id}"), response_model=LumaGeneration, status_extractor=lambda x: x.state, ) return IO.NodeOutput(await download_url_to_image_tensor(response_poll.assets.image)) @classmethod async def _convert_luma_refs(cls, luma_ref: LumaReferenceChain, max_refs: int): luma_urls = [] ref_count = 0 for ref in luma_ref.refs: download_urls = await upload_images_to_comfyapi(cls, ref.image, max_images=1) luma_urls.append(download_urls[0]) ref_count += 1 if ref_count >= max_refs: break return luma_ref.create_api_model(download_urls=luma_urls, max_refs=max_refs) @classmethod async def _convert_style_image(cls, style_image: torch.Tensor, weight: float): chain = LumaReferenceChain(first_ref=LumaReference(image=style_image, weight=weight)) return await cls._convert_luma_refs(chain, max_refs=1) class LumaImageModifyNode(IO.ComfyNode): @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id="LumaImageModifyNode", display_name="Luma Image to Image", category="api node/image/Luma", description="Modifies images synchronously based on prompt and aspect ratio.", inputs=[ IO.Image.Input( "image", ), IO.String.Input( "prompt", multiline=True, default="", tooltip="Prompt for the image generation", ), IO.Float.Input( "image_weight", default=0.1, min=0.0, max=0.98, step=0.01, tooltip="Weight of the image; the closer to 1.0, the less the image will be modified.", ), IO.Combo.Input( "model", options=LumaImageModel, ), IO.Int.Input( "seed", default=0, min=0, max=0xFFFFFFFFFFFFFFFF, control_after_generate=True, tooltip="Seed to determine if node should re-run; actual results are nondeterministic regardless of seed.", ), ], outputs=[IO.Image.Output()], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["model"]), expr=""" ( $m := widgets.model; $contains($m,"photon-flash-1") ? {"type":"usd","usd":0.0027} : $contains($m,"photon-1") ? {"type":"usd","usd":0.0104} : {"type":"usd","usd":0.0246} ) """, ), ) @classmethod async def execute( cls, prompt: str, model: str, image: torch.Tensor, image_weight: float, seed, ) -> IO.NodeOutput: download_urls = await upload_images_to_comfyapi(cls, image, max_images=1) image_url = download_urls[0] response_api = await sync_op( cls, ApiEndpoint(path="/proxy/luma/generations/image", method="POST"), response_model=LumaGeneration, data=LumaImageGenerationRequest( prompt=prompt, model=model, modify_image_ref=LumaModifyImageRef( url=image_url, weight=round(max(min(1.0 - image_weight, 0.98), 0.0), 2) ), ), ) response_poll = await poll_op( cls, ApiEndpoint(path=f"/proxy/luma/generations/{response_api.id}"), response_model=LumaGeneration, status_extractor=lambda x: x.state, ) return IO.NodeOutput(await download_url_to_image_tensor(response_poll.assets.image)) class LumaTextToVideoGenerationNode(IO.ComfyNode): @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id="LumaVideoNode", display_name="Luma Text to Video", category="api node/video/Luma", description="Generates videos synchronously based on prompt and output_size.", inputs=[ IO.String.Input( "prompt", multiline=True, default="", tooltip="Prompt for the video generation", ), IO.Combo.Input( "model", options=LumaVideoModel, ), IO.Combo.Input( "aspect_ratio", options=LumaAspectRatio, default=LumaAspectRatio.ratio_16_9, ), IO.Combo.Input( "resolution", options=LumaVideoOutputResolution, default=LumaVideoOutputResolution.res_540p, ), IO.Combo.Input( "duration", options=LumaVideoModelOutputDuration, ), IO.Boolean.Input( "loop", default=False, ), IO.Int.Input( "seed", default=0, min=0, max=0xFFFFFFFFFFFFFFFF, control_after_generate=True, tooltip="Seed to determine if node should re-run; actual results are nondeterministic regardless of seed.", ), IO.Custom(LumaIO.LUMA_CONCEPTS).Input( "luma_concepts", tooltip="Optional Camera Concepts to dictate camera motion via the Luma Concepts node.", optional=True, ), ], outputs=[IO.Video.Output()], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=PRICE_BADGE_VIDEO, ) @classmethod async def execute( cls, prompt: str, model: str, aspect_ratio: str, resolution: str, duration: str, loop: bool, seed, luma_concepts: Optional[LumaConceptChain] = None, ) -> IO.NodeOutput: validate_string(prompt, strip_whitespace=False, min_length=3) duration = duration if model != LumaVideoModel.ray_1_6 else None resolution = resolution if model != LumaVideoModel.ray_1_6 else None response_api = await sync_op( cls, ApiEndpoint(path="/proxy/luma/generations", method="POST"), response_model=LumaGeneration, data=LumaGenerationRequest( prompt=prompt, model=model, resolution=resolution, aspect_ratio=aspect_ratio, duration=duration, loop=loop, concepts=luma_concepts.create_api_model() if luma_concepts else None, ), ) response_poll = await poll_op( cls, ApiEndpoint(path=f"/proxy/luma/generations/{response_api.id}"), response_model=LumaGeneration, status_extractor=lambda x: x.state, estimated_duration=LUMA_T2V_AVERAGE_DURATION, ) return IO.NodeOutput(await download_url_to_video_output(response_poll.assets.video)) class LumaImageToVideoGenerationNode(IO.ComfyNode): @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id="LumaImageToVideoNode", display_name="Luma Image to Video", category="api node/video/Luma", description="Generates videos synchronously based on prompt, input images, and output_size.", inputs=[ IO.String.Input( "prompt", multiline=True, default="", tooltip="Prompt for the video generation", ), IO.Combo.Input( "model", options=LumaVideoModel, ), # IO.Combo.Input( # "aspect_ratio", # options=[ratio.value for ratio in LumaAspectRatio], # default=LumaAspectRatio.ratio_16_9, # ), IO.Combo.Input( "resolution", options=LumaVideoOutputResolution, default=LumaVideoOutputResolution.res_540p, ), IO.Combo.Input( "duration", options=[dur.value for dur in LumaVideoModelOutputDuration], ), IO.Boolean.Input( "loop", default=False, ), IO.Int.Input( "seed", default=0, min=0, max=0xFFFFFFFFFFFFFFFF, control_after_generate=True, tooltip="Seed to determine if node should re-run; actual results are nondeterministic regardless of seed.", ), IO.Image.Input( "first_image", tooltip="First frame of generated video.", optional=True, ), IO.Image.Input( "last_image", tooltip="Last frame of generated video.", optional=True, ), IO.Custom(LumaIO.LUMA_CONCEPTS).Input( "luma_concepts", tooltip="Optional Camera Concepts to dictate camera motion via the Luma Concepts node.", optional=True, ), ], outputs=[IO.Video.Output()], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=PRICE_BADGE_VIDEO, ) @classmethod async def execute( cls, prompt: str, model: str, resolution: str, duration: str, loop: bool, seed, first_image: torch.Tensor = None, last_image: torch.Tensor = None, luma_concepts: LumaConceptChain = None, ) -> IO.NodeOutput: if first_image is None and last_image is None: raise Exception("At least one of first_image and last_image requires an input.") keyframes = await cls._convert_to_keyframes(first_image, last_image) duration = duration if model != LumaVideoModel.ray_1_6 else None resolution = resolution if model != LumaVideoModel.ray_1_6 else None response_api = await sync_op( cls, ApiEndpoint(path="/proxy/luma/generations", method="POST"), response_model=LumaGeneration, data=LumaGenerationRequest( prompt=prompt, model=model, aspect_ratio=LumaAspectRatio.ratio_16_9, # ignored, but still needed by the API for some reason resolution=resolution, duration=duration, loop=loop, keyframes=keyframes, concepts=luma_concepts.create_api_model() if luma_concepts else None, ), ) response_poll = await poll_op( cls, poll_endpoint=ApiEndpoint(path=f"/proxy/luma/generations/{response_api.id}"), response_model=LumaGeneration, status_extractor=lambda x: x.state, estimated_duration=LUMA_I2V_AVERAGE_DURATION, ) return IO.NodeOutput(await download_url_to_video_output(response_poll.assets.video)) @classmethod async def _convert_to_keyframes( cls, first_image: torch.Tensor = None, last_image: torch.Tensor = None, ): if first_image is None and last_image is None: return None frame0 = None frame1 = None if first_image is not None: download_urls = await upload_images_to_comfyapi(cls, first_image, max_images=1) frame0 = LumaImageReference(type="image", url=download_urls[0]) if last_image is not None: download_urls = await upload_images_to_comfyapi(cls, last_image, max_images=1) frame1 = LumaImageReference(type="image", url=download_urls[0]) return LumaKeyframes(frame0=frame0, frame1=frame1) PRICE_BADGE_VIDEO = IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["model", "resolution", "duration"]), expr=""" ( $p := { "ray-flash-2": { "5s": {"4k":3.13,"1080p":0.79,"720p":0.34,"540p":0.2}, "9s": {"4k":5.65,"1080p":1.42,"720p":0.61,"540p":0.36} }, "ray-2": { "5s": {"4k":9.11,"1080p":2.27,"720p":1.02,"540p":0.57}, "9s": {"4k":16.4,"1080p":4.1,"720p":1.83,"540p":1.03} } }; $m := widgets.model; $d := widgets.duration; $r := widgets.resolution; $modelKey := $contains($m,"ray-flash-2") ? "ray-flash-2" : $contains($m,"ray-2") ? "ray-2" : $contains($m,"ray-1-6") ? "ray-1-6" : "other"; $durKey := $contains($d,"5s") ? "5s" : $contains($d,"9s") ? "9s" : ""; $resKey := $contains($r,"4k") ? "4k" : $contains($r,"1080p") ? "1080p" : $contains($r,"720p") ? "720p" : $contains($r,"540p") ? "540p" : ""; $modelPrices := $lookup($p, $modelKey); $durPrices := $lookup($modelPrices, $durKey); $v := $lookup($durPrices, $resKey); $price := ($modelKey = "ray-1-6") ? 0.5 : ($modelKey = "other") ? 0.79 : ($exists($v) ? $v : 0.79); {"type":"usd","usd": $price} ) """, ) class LumaExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[IO.ComfyNode]]: return [ LumaImageGenerationNode, LumaImageModifyNode, LumaTextToVideoGenerationNode, LumaImageToVideoGenerationNode, LumaReferenceNode, LumaConceptsNode, ] async def comfy_entrypoint() -> LumaExtension: return LumaExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api_nodes/nodes_luma.py", "license": "GNU General Public License v3.0", "lines": 622, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_api_nodes/nodes_minimax.py
from typing import Optional import torch from typing_extensions import override from comfy_api.latest import IO, ComfyExtension from comfy_api_nodes.apis.minimax import ( MinimaxFileRetrieveResponse, MiniMaxModel, MinimaxTaskResultResponse, MinimaxVideoGenerationRequest, MinimaxVideoGenerationResponse, SubjectReferenceItem, ) from comfy_api_nodes.util import ( ApiEndpoint, download_url_to_video_output, poll_op, sync_op, upload_images_to_comfyapi, validate_string, ) I2V_AVERAGE_DURATION = 114 T2V_AVERAGE_DURATION = 234 async def _generate_mm_video( cls: type[IO.ComfyNode], *, prompt_text: str, seed: int, model: str, image: Optional[torch.Tensor] = None, # used for ImageToVideo subject: Optional[torch.Tensor] = None, # used for SubjectToVideo average_duration: Optional[int] = None, ) -> IO.NodeOutput: if image is None: validate_string(prompt_text, field_name="prompt_text") image_url = None if image is not None: image_url = (await upload_images_to_comfyapi(cls, image, max_images=1))[0] # TODO: figure out how to deal with subject properly, API returns invalid params when using S2V-01 model subject_reference = None if subject is not None: subject_url = (await upload_images_to_comfyapi(cls, subject, max_images=1))[0] subject_reference = [SubjectReferenceItem(image=subject_url)] response = await sync_op( cls, ApiEndpoint(path="/proxy/minimax/video_generation", method="POST"), response_model=MinimaxVideoGenerationResponse, data=MinimaxVideoGenerationRequest( model=MiniMaxModel(model), prompt=prompt_text, callback_url=None, first_frame_image=image_url, subject_reference=subject_reference, prompt_optimizer=None, ), ) task_id = response.task_id if not task_id: raise Exception(f"MiniMax generation failed: {response.base_resp}") task_result = await poll_op( cls, ApiEndpoint(path="/proxy/minimax/query/video_generation", query_params={"task_id": task_id}), response_model=MinimaxTaskResultResponse, status_extractor=lambda x: x.status.value, estimated_duration=average_duration, ) file_id = task_result.file_id if file_id is None: raise Exception("Request was not successful. Missing file ID.") file_result = await sync_op( cls, ApiEndpoint(path="/proxy/minimax/files/retrieve", query_params={"file_id": int(file_id)}), response_model=MinimaxFileRetrieveResponse, ) file_url = file_result.file.download_url if file_url is None: raise Exception(f"No video was found in the response. Full response: {file_result.model_dump()}") if file_result.file.backup_download_url: try: return IO.NodeOutput(await download_url_to_video_output(file_url, timeout=10, max_retries=2)) except Exception: # if we have a second URL to retrieve the result, try again using that one return IO.NodeOutput( await download_url_to_video_output(file_result.file.backup_download_url, max_retries=3) ) return IO.NodeOutput(await download_url_to_video_output(file_url)) class MinimaxTextToVideoNode(IO.ComfyNode): @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id="MinimaxTextToVideoNode", display_name="MiniMax Text to Video", category="api node/video/MiniMax", description="Generates videos synchronously based on a prompt, and optional parameters.", inputs=[ IO.String.Input( "prompt_text", multiline=True, default="", tooltip="Text prompt to guide the video generation", ), IO.Combo.Input( "model", options=["T2V-01", "T2V-01-Director"], default="T2V-01", tooltip="Model to use for video generation", ), IO.Int.Input( "seed", default=0, min=0, max=0xFFFFFFFFFFFFFFFF, step=1, control_after_generate=True, tooltip="The random seed used for creating the noise.", optional=True, ), ], outputs=[IO.Video.Output()], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.43}""", ), ) @classmethod async def execute( cls, prompt_text: str, model: str = "T2V-01", seed: int = 0, ) -> IO.NodeOutput: return await _generate_mm_video( cls, prompt_text=prompt_text, seed=seed, model=model, image=None, subject=None, average_duration=T2V_AVERAGE_DURATION, ) class MinimaxImageToVideoNode(IO.ComfyNode): @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id="MinimaxImageToVideoNode", display_name="MiniMax Image to Video", category="api node/video/MiniMax", description="Generates videos synchronously based on an image and prompt, and optional parameters.", inputs=[ IO.Image.Input( "image", tooltip="Image to use as first frame of video generation", ), IO.String.Input( "prompt_text", multiline=True, default="", tooltip="Text prompt to guide the video generation", ), IO.Combo.Input( "model", options=["I2V-01-Director", "I2V-01", "I2V-01-live"], default="I2V-01", tooltip="Model to use for video generation", ), IO.Int.Input( "seed", default=0, min=0, max=0xFFFFFFFFFFFFFFFF, step=1, control_after_generate=True, tooltip="The random seed used for creating the noise.", optional=True, ), ], outputs=[IO.Video.Output()], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.43}""", ), ) @classmethod async def execute( cls, image: torch.Tensor, prompt_text: str, model: str = "I2V-01", seed: int = 0, ) -> IO.NodeOutput: return await _generate_mm_video( cls, prompt_text=prompt_text, seed=seed, model=model, image=image, subject=None, average_duration=I2V_AVERAGE_DURATION, ) class MinimaxSubjectToVideoNode(IO.ComfyNode): @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id="MinimaxSubjectToVideoNode", display_name="MiniMax Subject to Video", category="api node/video/MiniMax", description="Generates videos synchronously based on an image and prompt, and optional parameters.", inputs=[ IO.Image.Input( "subject", tooltip="Image of subject to reference for video generation", ), IO.String.Input( "prompt_text", multiline=True, default="", tooltip="Text prompt to guide the video generation", ), IO.Combo.Input( "model", options=["S2V-01"], default="S2V-01", tooltip="Model to use for video generation", ), IO.Int.Input( "seed", default=0, min=0, max=0xFFFFFFFFFFFFFFFF, step=1, control_after_generate=True, tooltip="The random seed used for creating the noise.", optional=True, ), ], outputs=[IO.Video.Output()], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, ) @classmethod async def execute( cls, subject: torch.Tensor, prompt_text: str, model: str = "S2V-01", seed: int = 0, ) -> IO.NodeOutput: return await _generate_mm_video( cls, prompt_text=prompt_text, seed=seed, model=model, image=None, subject=subject, average_duration=T2V_AVERAGE_DURATION, ) class MinimaxHailuoVideoNode(IO.ComfyNode): @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id="MinimaxHailuoVideoNode", display_name="MiniMax Hailuo Video", category="api node/video/MiniMax", description="Generates videos from prompt, with optional start frame using the new MiniMax Hailuo-02 model.", inputs=[ IO.String.Input( "prompt_text", multiline=True, default="", tooltip="Text prompt to guide the video generation.", ), IO.Int.Input( "seed", default=0, min=0, max=0xFFFFFFFFFFFFFFFF, step=1, control_after_generate=True, tooltip="The random seed used for creating the noise.", optional=True, ), IO.Image.Input( "first_frame_image", tooltip="Optional image to use as the first frame to generate a video.", optional=True, ), IO.Boolean.Input( "prompt_optimizer", default=True, tooltip="Optimize prompt to improve generation quality when needed.", optional=True, ), IO.Combo.Input( "duration", options=[6, 10], default=6, tooltip="The length of the output video in seconds.", optional=True, ), IO.Combo.Input( "resolution", options=["768P", "1080P"], default="768P", tooltip="The dimensions of the video display. 1080p is 1920x1080, 768p is 1366x768.", optional=True, ), ], outputs=[IO.Video.Output()], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["resolution", "duration"]), expr=""" ( $prices := { "768p": {"6": 0.28, "10": 0.56}, "1080p": {"6": 0.49} }; $resPrices := $lookup($prices, $lowercase(widgets.resolution)); $price := $lookup($resPrices, $string(widgets.duration)); {"type":"usd","usd": $price ? $price : 0.43} ) """, ), ) @classmethod async def execute( cls, prompt_text: str, seed: int = 0, first_frame_image: Optional[torch.Tensor] = None, # used for ImageToVideo prompt_optimizer: bool = True, duration: int = 6, resolution: str = "768P", model: str = "MiniMax-Hailuo-02", ) -> IO.NodeOutput: if first_frame_image is None: validate_string(prompt_text, field_name="prompt_text") if model == "MiniMax-Hailuo-02" and resolution.upper() == "1080P" and duration != 6: raise Exception( "When model is MiniMax-Hailuo-02 and resolution is 1080P, duration is limited to 6 seconds." ) # upload image, if passed in image_url = None if first_frame_image is not None: image_url = (await upload_images_to_comfyapi(cls, first_frame_image, max_images=1))[0] response = await sync_op( cls, ApiEndpoint(path="/proxy/minimax/video_generation", method="POST"), response_model=MinimaxVideoGenerationResponse, data=MinimaxVideoGenerationRequest( model=MiniMaxModel(model), prompt=prompt_text, callback_url=None, first_frame_image=image_url, prompt_optimizer=prompt_optimizer, duration=duration, resolution=resolution, ), ) task_id = response.task_id if not task_id: raise Exception(f"MiniMax generation failed: {response.base_resp}") average_duration = 120 if resolution == "768P" else 240 task_result = await poll_op( cls, ApiEndpoint(path="/proxy/minimax/query/video_generation", query_params={"task_id": task_id}), response_model=MinimaxTaskResultResponse, status_extractor=lambda x: x.status.value, estimated_duration=average_duration, ) file_id = task_result.file_id if file_id is None: raise Exception("Request was not successful. Missing file ID.") file_result = await sync_op( cls, ApiEndpoint(path="/proxy/minimax/files/retrieve", query_params={"file_id": int(file_id)}), response_model=MinimaxFileRetrieveResponse, ) file_url = file_result.file.download_url if file_url is None: raise Exception(f"No video was found in the response. Full response: {file_result.model_dump()}") if file_result.file.backup_download_url: try: return IO.NodeOutput(await download_url_to_video_output(file_url, timeout=10, max_retries=2)) except Exception: # if we have a second URL to retrieve the result, try again using that one return IO.NodeOutput( await download_url_to_video_output(file_result.file.backup_download_url, max_retries=3) ) return IO.NodeOutput(await download_url_to_video_output(file_url)) class MinimaxExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[IO.ComfyNode]]: return [ MinimaxTextToVideoNode, MinimaxImageToVideoNode, # MinimaxSubjectToVideoNode, MinimaxHailuoVideoNode, ] async def comfy_entrypoint() -> MinimaxExtension: return MinimaxExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api_nodes/nodes_minimax.py", "license": "GNU General Public License v3.0", "lines": 417, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_api_nodes/nodes_openai.py
import base64 import os from enum import Enum from io import BytesIO import numpy as np import torch from PIL import Image from typing_extensions import override import folder_paths from comfy_api.latest import IO, ComfyExtension, Input from comfy_api_nodes.apis.openai import ( InputFileContent, InputImageContent, InputMessage, InputTextContent, ModelResponseProperties, OpenAICreateResponse, OpenAIImageEditRequest, OpenAIImageGenerationRequest, OpenAIImageGenerationResponse, OpenAIResponse, OutputContent, ) from comfy_api_nodes.util import ( ApiEndpoint, download_url_to_bytesio, downscale_image_tensor, poll_op, sync_op, tensor_to_base64_string, text_filepath_to_data_uri, validate_string, ) RESPONSES_ENDPOINT = "/proxy/openai/v1/responses" STARTING_POINT_ID_PATTERN = r"<starting_point_id:(.*)>" class SupportedOpenAIModel(str, Enum): o4_mini = "o4-mini" o1 = "o1" o3 = "o3" o1_pro = "o1-pro" gpt_4_1 = "gpt-4.1" gpt_4_1_mini = "gpt-4.1-mini" gpt_4_1_nano = "gpt-4.1-nano" gpt_5 = "gpt-5" gpt_5_mini = "gpt-5-mini" gpt_5_nano = "gpt-5-nano" async def validate_and_cast_response(response, timeout: int = None) -> torch.Tensor: """Validates and casts a response to a torch.Tensor. Args: response: The response to validate and cast. timeout: Request timeout in seconds. Defaults to None (no timeout). Returns: A torch.Tensor representing the image (1, H, W, C). Raises: ValueError: If the response is not valid. """ # validate raw JSON response data = response.data if not data or len(data) == 0: raise ValueError("No images returned from API endpoint") # Initialize list to store image tensors image_tensors: list[torch.Tensor] = [] # Process each image in the data array for img_data in data: if img_data.b64_json: img_io = BytesIO(base64.b64decode(img_data.b64_json)) elif img_data.url: img_io = BytesIO() await download_url_to_bytesio(img_data.url, img_io, timeout=timeout) else: raise ValueError("Invalid image payload – neither URL nor base64 data present.") pil_img = Image.open(img_io).convert("RGBA") arr = np.asarray(pil_img).astype(np.float32) / 255.0 image_tensors.append(torch.from_numpy(arr)) return torch.stack(image_tensors, dim=0) class OpenAIDalle2(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="OpenAIDalle2", display_name="OpenAI DALL·E 2", category="api node/image/OpenAI", description="Generates images synchronously via OpenAI's DALL·E 2 endpoint.", inputs=[ IO.String.Input( "prompt", default="", multiline=True, tooltip="Text prompt for DALL·E", ), IO.Int.Input( "seed", default=0, min=0, max=2**31 - 1, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, tooltip="not implemented yet in backend", optional=True, ), IO.Combo.Input( "size", default="1024x1024", options=["256x256", "512x512", "1024x1024"], tooltip="Image size", optional=True, ), IO.Int.Input( "n", default=1, min=1, max=8, step=1, tooltip="How many images to generate", display_mode=IO.NumberDisplay.number, optional=True, ), IO.Image.Input( "image", tooltip="Optional reference image for image editing.", optional=True, ), IO.Mask.Input( "mask", tooltip="Optional mask for inpainting (white areas will be replaced)", optional=True, ), ], outputs=[ IO.Image.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["size", "n"]), expr=""" ( $size := widgets.size; $nRaw := widgets.n; $n := ($nRaw != null and $nRaw != 0) ? $nRaw : 1; $base := $contains($size, "256x256") ? 0.016 : $contains($size, "512x512") ? 0.018 : 0.02; {"type":"usd","usd": $round($base * $n, 3)} ) """, ), ) @classmethod async def execute( cls, prompt, seed=0, image=None, mask=None, n=1, size="1024x1024", ) -> IO.NodeOutput: validate_string(prompt, strip_whitespace=False) model = "dall-e-2" path = "/proxy/openai/images/generations" content_type = "application/json" request_class = OpenAIImageGenerationRequest img_binary = None if image is not None and mask is not None: path = "/proxy/openai/images/edits" content_type = "multipart/form-data" request_class = OpenAIImageEditRequest input_tensor = image.squeeze().cpu() height, width, channels = input_tensor.shape rgba_tensor = torch.ones(height, width, 4, device="cpu") rgba_tensor[:, :, :channels] = input_tensor if mask.shape[1:] != image.shape[1:-1]: raise Exception("Mask and Image must be the same size") rgba_tensor[:, :, 3] = 1 - mask.squeeze().cpu() rgba_tensor = downscale_image_tensor(rgba_tensor.unsqueeze(0)).squeeze() image_np = (rgba_tensor.numpy() * 255).astype(np.uint8) img = Image.fromarray(image_np) img_byte_arr = BytesIO() img.save(img_byte_arr, format="PNG") img_byte_arr.seek(0) img_binary = img_byte_arr # .getvalue() img_binary.name = "image.png" elif image is not None or mask is not None: raise Exception("Dall-E 2 image editing requires an image AND a mask") response = await sync_op( cls, ApiEndpoint(path=path, method="POST"), response_model=OpenAIImageGenerationResponse, data=request_class( model=model, prompt=prompt, n=n, size=size, seed=seed, ), files=( { "image": ("image.png", img_binary, "image/png"), } if img_binary else None ), content_type=content_type, ) return IO.NodeOutput(await validate_and_cast_response(response)) class OpenAIDalle3(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="OpenAIDalle3", display_name="OpenAI DALL·E 3", category="api node/image/OpenAI", description="Generates images synchronously via OpenAI's DALL·E 3 endpoint.", inputs=[ IO.String.Input( "prompt", default="", multiline=True, tooltip="Text prompt for DALL·E", ), IO.Int.Input( "seed", default=0, min=0, max=2**31 - 1, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, tooltip="not implemented yet in backend", optional=True, ), IO.Combo.Input( "quality", default="standard", options=["standard", "hd"], tooltip="Image quality", optional=True, ), IO.Combo.Input( "style", default="natural", options=["natural", "vivid"], tooltip="Vivid causes the model to lean towards generating hyper-real and dramatic images. Natural causes the model to produce more natural, less hyper-real looking images.", optional=True, ), IO.Combo.Input( "size", default="1024x1024", options=["1024x1024", "1024x1792", "1792x1024"], tooltip="Image size", optional=True, ), ], outputs=[ IO.Image.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["size", "quality"]), expr=""" ( $size := widgets.size; $q := widgets.quality; $hd := $contains($q, "hd"); $price := $contains($size, "1024x1024") ? ($hd ? 0.08 : 0.04) : (($contains($size, "1792x1024") or $contains($size, "1024x1792")) ? ($hd ? 0.12 : 0.08) : 0.04); {"type":"usd","usd": $price} ) """, ), ) @classmethod async def execute( cls, prompt, seed=0, style="natural", quality="standard", size="1024x1024", ) -> IO.NodeOutput: validate_string(prompt, strip_whitespace=False) model = "dall-e-3" # build the operation response = await sync_op( cls, ApiEndpoint(path="/proxy/openai/images/generations", method="POST"), response_model=OpenAIImageGenerationResponse, data=OpenAIImageGenerationRequest( model=model, prompt=prompt, quality=quality, size=size, style=style, seed=seed, ), ) return IO.NodeOutput(await validate_and_cast_response(response)) def calculate_tokens_price_image_1(response: OpenAIImageGenerationResponse) -> float | None: # https://platform.openai.com/docs/pricing return ((response.usage.input_tokens * 10.0) + (response.usage.output_tokens * 40.0)) / 1_000_000.0 def calculate_tokens_price_image_1_5(response: OpenAIImageGenerationResponse) -> float | None: return ((response.usage.input_tokens * 8.0) + (response.usage.output_tokens * 32.0)) / 1_000_000.0 class OpenAIGPTImage1(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="OpenAIGPTImage1", display_name="OpenAI GPT Image 1.5", category="api node/image/OpenAI", description="Generates images synchronously via OpenAI's GPT Image endpoint.", inputs=[ IO.String.Input( "prompt", default="", multiline=True, tooltip="Text prompt for GPT Image", ), IO.Int.Input( "seed", default=0, min=0, max=2**31 - 1, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, tooltip="not implemented yet in backend", optional=True, ), IO.Combo.Input( "quality", default="low", options=["low", "medium", "high"], tooltip="Image quality, affects cost and generation time.", optional=True, ), IO.Combo.Input( "background", default="auto", options=["auto", "opaque", "transparent"], tooltip="Return image with or without background", optional=True, ), IO.Combo.Input( "size", default="auto", options=["auto", "1024x1024", "1024x1536", "1536x1024"], tooltip="Image size", optional=True, ), IO.Int.Input( "n", default=1, min=1, max=8, step=1, tooltip="How many images to generate", display_mode=IO.NumberDisplay.number, optional=True, ), IO.Image.Input( "image", tooltip="Optional reference image for image editing.", optional=True, ), IO.Mask.Input( "mask", tooltip="Optional mask for inpainting (white areas will be replaced)", optional=True, ), IO.Combo.Input( "model", options=["gpt-image-1", "gpt-image-1.5"], default="gpt-image-1.5", optional=True, ), ], outputs=[ IO.Image.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["quality", "n"]), expr=""" ( $ranges := { "low": [0.011, 0.02], "medium": [0.046, 0.07], "high": [0.167, 0.3] }; $range := $lookup($ranges, widgets.quality); $n := widgets.n; ($n = 1) ? {"type":"range_usd","min_usd": $range[0], "max_usd": $range[1]} : { "type":"range_usd", "min_usd": $range[0], "max_usd": $range[1], "format": { "suffix": " x " & $string($n) & "/Run" } } ) """, ), ) @classmethod async def execute( cls, prompt: str, seed: int = 0, quality: str = "low", background: str = "opaque", image: Input.Image | None = None, mask: Input.Image | None = None, n: int = 1, size: str = "1024x1024", model: str = "gpt-image-1", ) -> IO.NodeOutput: validate_string(prompt, strip_whitespace=False) if mask is not None and image is None: raise ValueError("Cannot use a mask without an input image") if model == "gpt-image-1": price_extractor = calculate_tokens_price_image_1 elif model == "gpt-image-1.5": price_extractor = calculate_tokens_price_image_1_5 else: raise ValueError(f"Unknown model: {model}") if image is not None: files = [] batch_size = image.shape[0] for i in range(batch_size): single_image = image[i : i + 1] scaled_image = downscale_image_tensor(single_image, total_pixels=2048 * 2048).squeeze() image_np = (scaled_image.numpy() * 255).astype(np.uint8) img = Image.fromarray(image_np) img_byte_arr = BytesIO() img.save(img_byte_arr, format="PNG") img_byte_arr.seek(0) if batch_size == 1: files.append(("image", (f"image_{i}.png", img_byte_arr, "image/png"))) else: files.append(("image[]", (f"image_{i}.png", img_byte_arr, "image/png"))) if mask is not None: if image.shape[0] != 1: raise Exception("Cannot use a mask with multiple image") if mask.shape[1:] != image.shape[1:-1]: raise Exception("Mask and Image must be the same size") _, height, width = mask.shape rgba_mask = torch.zeros(height, width, 4, device="cpu") rgba_mask[:, :, 3] = 1 - mask.squeeze().cpu() scaled_mask = downscale_image_tensor(rgba_mask.unsqueeze(0), total_pixels=2048 * 2048).squeeze() mask_np = (scaled_mask.numpy() * 255).astype(np.uint8) mask_img = Image.fromarray(mask_np) mask_img_byte_arr = BytesIO() mask_img.save(mask_img_byte_arr, format="PNG") mask_img_byte_arr.seek(0) files.append(("mask", ("mask.png", mask_img_byte_arr, "image/png"))) response = await sync_op( cls, ApiEndpoint(path="/proxy/openai/images/edits", method="POST"), response_model=OpenAIImageGenerationResponse, data=OpenAIImageEditRequest( model=model, prompt=prompt, quality=quality, background=background, n=n, seed=seed, size=size, moderation="low", ), content_type="multipart/form-data", files=files, price_extractor=price_extractor, ) else: response = await sync_op( cls, ApiEndpoint(path="/proxy/openai/images/generations", method="POST"), response_model=OpenAIImageGenerationResponse, data=OpenAIImageGenerationRequest( model=model, prompt=prompt, quality=quality, background=background, n=n, seed=seed, size=size, moderation="low", ), price_extractor=price_extractor, ) return IO.NodeOutput(await validate_and_cast_response(response)) class OpenAIChatNode(IO.ComfyNode): """ Node to generate text responses from an OpenAI model. """ @classmethod def define_schema(cls): return IO.Schema( node_id="OpenAIChatNode", display_name="OpenAI ChatGPT", category="api node/text/OpenAI", essentials_category="Text Generation", description="Generate text responses from an OpenAI model.", inputs=[ IO.String.Input( "prompt", default="", multiline=True, tooltip="Text inputs to the model, used to generate a response.", ), IO.Boolean.Input( "persist_context", default=False, tooltip="This parameter is deprecated and has no effect.", advanced=True, ), IO.Combo.Input( "model", options=SupportedOpenAIModel, tooltip="The model used to generate the response", ), IO.Image.Input( "images", tooltip="Optional image(s) to use as context for the model. To include multiple images, you can use the Batch Images node.", optional=True, ), IO.Custom("OPENAI_INPUT_FILES").Input( "files", optional=True, tooltip="Optional file(s) to use as context for the model. Accepts inputs from the OpenAI Chat Input Files node.", ), IO.Custom("OPENAI_CHAT_CONFIG").Input( "advanced_options", optional=True, tooltip="Optional configuration for the model. Accepts inputs from the OpenAI Chat Advanced Options node.", ), ], outputs=[ IO.String.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["model"]), expr=""" ( $m := widgets.model; $contains($m, "o4-mini") ? { "type": "list_usd", "usd": [0.0011, 0.0044], "format": { "approximate": true, "separator": "-", "suffix": " per 1K tokens" } } : $contains($m, "o1-pro") ? { "type": "list_usd", "usd": [0.15, 0.6], "format": { "approximate": true, "separator": "-", "suffix": " per 1K tokens" } } : $contains($m, "o1") ? { "type": "list_usd", "usd": [0.015, 0.06], "format": { "approximate": true, "separator": "-", "suffix": " per 1K tokens" } } : $contains($m, "o3-mini") ? { "type": "list_usd", "usd": [0.0011, 0.0044], "format": { "approximate": true, "separator": "-", "suffix": " per 1K tokens" } } : $contains($m, "o3") ? { "type": "list_usd", "usd": [0.01, 0.04], "format": { "approximate": true, "separator": "-", "suffix": " per 1K tokens" } } : $contains($m, "gpt-4.1-nano") ? { "type": "list_usd", "usd": [0.0001, 0.0004], "format": { "approximate": true, "separator": "-", "suffix": " per 1K tokens" } } : $contains($m, "gpt-4.1-mini") ? { "type": "list_usd", "usd": [0.0004, 0.0016], "format": { "approximate": true, "separator": "-", "suffix": " per 1K tokens" } } : $contains($m, "gpt-4.1") ? { "type": "list_usd", "usd": [0.002, 0.008], "format": { "approximate": true, "separator": "-", "suffix": " per 1K tokens" } } : $contains($m, "gpt-5-nano") ? { "type": "list_usd", "usd": [0.00005, 0.0004], "format": { "approximate": true, "separator": "-", "suffix": " per 1K tokens" } } : $contains($m, "gpt-5-mini") ? { "type": "list_usd", "usd": [0.00025, 0.002], "format": { "approximate": true, "separator": "-", "suffix": " per 1K tokens" } } : $contains($m, "gpt-5") ? { "type": "list_usd", "usd": [0.00125, 0.01], "format": { "approximate": true, "separator": "-", "suffix": " per 1K tokens" } } : {"type": "text", "text": "Token-based"} ) """, ), ) @classmethod def get_message_content_from_response(cls, response: OpenAIResponse) -> list[OutputContent]: """Extract message content from the API response.""" for output in response.output: if output.type == "message": return output.content raise TypeError("No output message found in response") @classmethod def get_text_from_message_content(cls, message_content: list[OutputContent]) -> str: """Extract text content from message content.""" for content_item in message_content: if content_item.type == "output_text": return str(content_item.text) return "No text output found in response" @classmethod def tensor_to_input_image_content(cls, image: torch.Tensor, detail_level: str = "auto") -> InputImageContent: """Convert a tensor to an input image content object.""" return InputImageContent( detail=detail_level, image_url=f"data:image/png;base64,{tensor_to_base64_string(image)}", type="input_image", ) @classmethod def create_input_message_contents( cls, prompt: str, image: torch.Tensor | None = None, files: list[InputFileContent] | None = None, ) -> list[InputTextContent | InputImageContent | InputFileContent]: """Create a list of input message contents from prompt and optional image.""" content_list: list[InputTextContent | InputImageContent | InputFileContent] = [ InputTextContent(text=prompt, type="input_text"), ] if image is not None: for i in range(image.shape[0]): content_list.append( InputImageContent( detail="auto", image_url=f"data:image/png;base64,{tensor_to_base64_string(image[i].unsqueeze(0))}", type="input_image", ) ) if files is not None: content_list.extend(files) return content_list @classmethod async def execute( cls, prompt: str, persist_context: bool = False, model: SupportedOpenAIModel = SupportedOpenAIModel.gpt_5.value, images: torch.Tensor | None = None, files: list[InputFileContent] | None = None, advanced_options: ModelResponseProperties | None = None, ) -> IO.NodeOutput: validate_string(prompt, strip_whitespace=False) # Create response create_response = await sync_op( cls, ApiEndpoint(path=RESPONSES_ENDPOINT, method="POST"), response_model=OpenAIResponse, data=OpenAICreateResponse( input=[ InputMessage( content=cls.create_input_message_contents(prompt, images, files), role="user", ), ], store=True, stream=False, model=model, previous_response_id=None, **(advanced_options.model_dump(exclude_none=True) if advanced_options else {}), ), ) response_id = create_response.id # Get result output result_response = await poll_op( cls, ApiEndpoint(path=f"{RESPONSES_ENDPOINT}/{response_id}"), response_model=OpenAIResponse, status_extractor=lambda response: response.status, completed_statuses=["incomplete", "completed"], ) return IO.NodeOutput(cls.get_text_from_message_content(cls.get_message_content_from_response(result_response))) class OpenAIInputFiles(IO.ComfyNode): """ Loads and formats input files for OpenAI API. """ @classmethod def define_schema(cls): """ For details about the supported file input types, see: https://platform.openai.com/docs/guides/pdf-files?api-mode=responses """ input_dir = folder_paths.get_input_directory() input_files = [ f for f in os.scandir(input_dir) if f.is_file() and (f.name.endswith(".txt") or f.name.endswith(".pdf")) and f.stat().st_size < 32 * 1024 * 1024 ] input_files = sorted(input_files, key=lambda x: x.name) input_files = [f.name for f in input_files] return IO.Schema( node_id="OpenAIInputFiles", display_name="OpenAI ChatGPT Input Files", category="api node/text/OpenAI", description="Loads and prepares input files (text, pdf, etc.) to include as inputs for the OpenAI Chat Node. The files will be read by the OpenAI model when generating a response. 🛈 TIP: Can be chained together with other OpenAI Input File nodes.", inputs=[ IO.Combo.Input( "file", options=input_files, default=input_files[0] if input_files else None, tooltip="Input files to include as context for the model. Only accepts text (.txt) and PDF (.pdf) files for now.", ), IO.Custom("OPENAI_INPUT_FILES").Input( "OPENAI_INPUT_FILES", tooltip="An optional additional file(s) to batch together with the file loaded from this node. Allows chaining of input files so that a single message can include multiple input files.", optional=True, ), ], outputs=[ IO.Custom("OPENAI_INPUT_FILES").Output(), ], ) @classmethod def create_input_file_content(cls, file_path: str) -> InputFileContent: return InputFileContent( file_data=text_filepath_to_data_uri(file_path), filename=os.path.basename(file_path), type="input_file", ) @classmethod def execute(cls, file: str, OPENAI_INPUT_FILES: list[InputFileContent] = []) -> IO.NodeOutput: """ Loads and formats input files for OpenAI API. """ file_path = folder_paths.get_annotated_filepath(file) input_file_content = cls.create_input_file_content(file_path) files = [input_file_content] + OPENAI_INPUT_FILES return IO.NodeOutput(files) class OpenAIChatConfig(IO.ComfyNode): """Allows setting additional configuration for the OpenAI Chat Node.""" @classmethod def define_schema(cls): return IO.Schema( node_id="OpenAIChatConfig", display_name="OpenAI ChatGPT Advanced Options", category="api node/text/OpenAI", description="Allows specifying advanced configuration options for the OpenAI Chat Nodes.", inputs=[ IO.Combo.Input( "truncation", options=["auto", "disabled"], default="auto", tooltip="The truncation strategy to use for the model response. auto: If the context of this response and previous ones exceeds the model's context window size, the model will truncate the response to fit the context window by dropping input items in the middle of the conversation.disabled: If a model response will exceed the context window size for a model, the request will fail with a 400 error", advanced=True, ), IO.Int.Input( "max_output_tokens", min=16, default=4096, max=16384, tooltip="An upper bound for the number of tokens that can be generated for a response, including visible output tokens", optional=True, advanced=True, ), IO.String.Input( "instructions", multiline=True, optional=True, tooltip="Instructions for the model on how to generate the response", ), ], outputs=[ IO.Custom("OPENAI_CHAT_CONFIG").Output(), ], ) @classmethod def execute( cls, truncation: bool, instructions: str | None = None, max_output_tokens: int | None = None, ) -> IO.NodeOutput: """ Configure advanced options for the OpenAI Chat Node. Note: While `top_p` and `temperature` are listed as properties in the spec, they are not supported for all models (e.g., o4-mini). They are not exposed as inputs at all to avoid having to manually remove depending on model choice. """ return IO.NodeOutput( ModelResponseProperties( instructions=instructions, truncation=truncation, max_output_tokens=max_output_tokens, ) ) class OpenAIExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[IO.ComfyNode]]: return [ OpenAIDalle2, OpenAIDalle3, OpenAIGPTImage1, OpenAIChatNode, OpenAIInputFiles, OpenAIChatConfig, ] async def comfy_entrypoint() -> OpenAIExtension: return OpenAIExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api_nodes/nodes_openai.py", "license": "GNU General Public License v3.0", "lines": 847, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_api_nodes/nodes_pixverse.py
import torch from typing_extensions import override from comfy_api.latest import IO, ComfyExtension from comfy_api_nodes.apis.pixverse import ( PixverseTextVideoRequest, PixverseImageVideoRequest, PixverseTransitionVideoRequest, PixverseImageUploadResponse, PixverseVideoResponse, PixverseGenerationStatusResponse, PixverseAspectRatio, PixverseQuality, PixverseDuration, PixverseMotionMode, PixverseStatus, PixverseIO, pixverse_templates, ) from comfy_api_nodes.util import ( ApiEndpoint, download_url_to_video_output, poll_op, sync_op, tensor_to_bytesio, validate_string, ) AVERAGE_DURATION_T2V = 32 AVERAGE_DURATION_I2V = 30 AVERAGE_DURATION_T2T = 52 async def upload_image_to_pixverse(cls: type[IO.ComfyNode], image: torch.Tensor): response_upload = await sync_op( cls, ApiEndpoint(path="/proxy/pixverse/image/upload", method="POST"), response_model=PixverseImageUploadResponse, files={"image": tensor_to_bytesio(image)}, content_type="multipart/form-data", ) if response_upload.Resp is None: raise Exception(f"PixVerse image upload request failed: '{response_upload.ErrMsg}'") return response_upload.Resp.img_id class PixverseTemplateNode(IO.ComfyNode): """ Select template for PixVerse Video generation. """ @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id="PixverseTemplateNode", display_name="PixVerse Template", category="api node/video/PixVerse", inputs=[ IO.Combo.Input("template", options=list(pixverse_templates.keys())), ], outputs=[IO.Custom(PixverseIO.TEMPLATE).Output(display_name="pixverse_template")], ) @classmethod def execute(cls, template: str) -> IO.NodeOutput: template_id = pixverse_templates.get(template, None) if template_id is None: raise Exception(f"Template '{template}' is not recognized.") return IO.NodeOutput(template_id) class PixverseTextToVideoNode(IO.ComfyNode): @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id="PixverseTextToVideoNode", display_name="PixVerse Text to Video", category="api node/video/PixVerse", description="Generates videos based on prompt and output_size.", inputs=[ IO.String.Input( "prompt", multiline=True, default="", tooltip="Prompt for the video generation", ), IO.Combo.Input( "aspect_ratio", options=PixverseAspectRatio, ), IO.Combo.Input( "quality", options=PixverseQuality, default=PixverseQuality.res_540p, ), IO.Combo.Input( "duration_seconds", options=PixverseDuration, ), IO.Combo.Input( "motion_mode", options=PixverseMotionMode, ), IO.Int.Input( "seed", default=0, min=0, max=2147483647, control_after_generate=True, tooltip="Seed for video generation.", ), IO.String.Input( "negative_prompt", default="", multiline=True, tooltip="An optional text description of undesired elements on an image.", optional=True, ), IO.Custom(PixverseIO.TEMPLATE).Input( "pixverse_template", tooltip="An optional template to influence style of generation, created by the PixVerse Template node.", optional=True, ), ], outputs=[IO.Video.Output()], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=PRICE_BADGE_VIDEO, ) @classmethod async def execute( cls, prompt: str, aspect_ratio: str, quality: str, duration_seconds: int, motion_mode: str, seed, negative_prompt: str = None, pixverse_template: int = None, ) -> IO.NodeOutput: validate_string(prompt, strip_whitespace=False, min_length=1) # 1080p is limited to 5 seconds duration # only normal motion_mode supported for 1080p or for non-5 second duration if quality == PixverseQuality.res_1080p: motion_mode = PixverseMotionMode.normal duration_seconds = PixverseDuration.dur_5 elif duration_seconds != PixverseDuration.dur_5: motion_mode = PixverseMotionMode.normal response_api = await sync_op( cls, ApiEndpoint(path="/proxy/pixverse/video/text/generate", method="POST"), response_model=PixverseVideoResponse, data=PixverseTextVideoRequest( prompt=prompt, aspect_ratio=aspect_ratio, quality=quality, duration=duration_seconds, motion_mode=motion_mode, negative_prompt=negative_prompt if negative_prompt else None, template_id=pixverse_template, seed=seed, ), ) if response_api.Resp is None: raise Exception(f"PixVerse request failed: '{response_api.ErrMsg}'") response_poll = await poll_op( cls, ApiEndpoint(path=f"/proxy/pixverse/video/result/{response_api.Resp.video_id}"), response_model=PixverseGenerationStatusResponse, completed_statuses=[PixverseStatus.successful], failed_statuses=[ PixverseStatus.contents_moderation, PixverseStatus.failed, PixverseStatus.deleted, ], status_extractor=lambda x: x.Resp.status, estimated_duration=AVERAGE_DURATION_T2V, ) return IO.NodeOutput(await download_url_to_video_output(response_poll.Resp.url)) class PixverseImageToVideoNode(IO.ComfyNode): @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id="PixverseImageToVideoNode", display_name="PixVerse Image to Video", category="api node/video/PixVerse", description="Generates videos based on prompt and output_size.", inputs=[ IO.Image.Input("image"), IO.String.Input( "prompt", multiline=True, default="", tooltip="Prompt for the video generation", ), IO.Combo.Input( "quality", options=PixverseQuality, default=PixverseQuality.res_540p, ), IO.Combo.Input( "duration_seconds", options=PixverseDuration, ), IO.Combo.Input( "motion_mode", options=PixverseMotionMode, ), IO.Int.Input( "seed", default=0, min=0, max=2147483647, control_after_generate=True, tooltip="Seed for video generation.", ), IO.String.Input( "negative_prompt", default="", multiline=True, tooltip="An optional text description of undesired elements on an image.", optional=True, ), IO.Custom(PixverseIO.TEMPLATE).Input( "pixverse_template", tooltip="An optional template to influence style of generation, created by the PixVerse Template node.", optional=True, ), ], outputs=[IO.Video.Output()], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=PRICE_BADGE_VIDEO, ) @classmethod async def execute( cls, image: torch.Tensor, prompt: str, quality: str, duration_seconds: int, motion_mode: str, seed, negative_prompt: str = None, pixverse_template: int = None, ) -> IO.NodeOutput: validate_string(prompt, strip_whitespace=False) img_id = await upload_image_to_pixverse(cls, image) # 1080p is limited to 5 seconds duration # only normal motion_mode supported for 1080p or for non-5 second duration if quality == PixverseQuality.res_1080p: motion_mode = PixverseMotionMode.normal duration_seconds = PixverseDuration.dur_5 elif duration_seconds != PixverseDuration.dur_5: motion_mode = PixverseMotionMode.normal response_api = await sync_op( cls, ApiEndpoint(path="/proxy/pixverse/video/img/generate", method="POST"), response_model=PixverseVideoResponse, data=PixverseImageVideoRequest( img_id=img_id, prompt=prompt, quality=quality, duration=duration_seconds, motion_mode=motion_mode, negative_prompt=negative_prompt if negative_prompt else None, template_id=pixverse_template, seed=seed, ), ) if response_api.Resp is None: raise Exception(f"PixVerse request failed: '{response_api.ErrMsg}'") response_poll = await poll_op( cls, ApiEndpoint(path=f"/proxy/pixverse/video/result/{response_api.Resp.video_id}"), response_model=PixverseGenerationStatusResponse, completed_statuses=[PixverseStatus.successful], failed_statuses=[ PixverseStatus.contents_moderation, PixverseStatus.failed, PixverseStatus.deleted, ], status_extractor=lambda x: x.Resp.status, estimated_duration=AVERAGE_DURATION_I2V, ) return IO.NodeOutput(await download_url_to_video_output(response_poll.Resp.url)) class PixverseTransitionVideoNode(IO.ComfyNode): @classmethod def define_schema(cls) -> IO.Schema: return IO.Schema( node_id="PixverseTransitionVideoNode", display_name="PixVerse Transition Video", category="api node/video/PixVerse", description="Generates videos based on prompt and output_size.", inputs=[ IO.Image.Input("first_frame"), IO.Image.Input("last_frame"), IO.String.Input( "prompt", multiline=True, default="", tooltip="Prompt for the video generation", ), IO.Combo.Input( "quality", options=PixverseQuality, default=PixverseQuality.res_540p, ), IO.Combo.Input( "duration_seconds", options=PixverseDuration, ), IO.Combo.Input( "motion_mode", options=PixverseMotionMode, ), IO.Int.Input( "seed", default=0, min=0, max=2147483647, control_after_generate=True, tooltip="Seed for video generation.", ), IO.String.Input( "negative_prompt", default="", multiline=True, tooltip="An optional text description of undesired elements on an image.", optional=True, ), ], outputs=[IO.Video.Output()], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=PRICE_BADGE_VIDEO, ) @classmethod async def execute( cls, first_frame: torch.Tensor, last_frame: torch.Tensor, prompt: str, quality: str, duration_seconds: int, motion_mode: str, seed, negative_prompt: str = None, ) -> IO.NodeOutput: validate_string(prompt, strip_whitespace=False) first_frame_id = await upload_image_to_pixverse(cls, first_frame) last_frame_id = await upload_image_to_pixverse(cls, last_frame) # 1080p is limited to 5 seconds duration # only normal motion_mode supported for 1080p or for non-5 second duration if quality == PixverseQuality.res_1080p: motion_mode = PixverseMotionMode.normal duration_seconds = PixverseDuration.dur_5 elif duration_seconds != PixverseDuration.dur_5: motion_mode = PixverseMotionMode.normal response_api = await sync_op( cls, ApiEndpoint(path="/proxy/pixverse/video/transition/generate", method="POST"), response_model=PixverseVideoResponse, data=PixverseTransitionVideoRequest( first_frame_img=first_frame_id, last_frame_img=last_frame_id, prompt=prompt, quality=quality, duration=duration_seconds, motion_mode=motion_mode, negative_prompt=negative_prompt if negative_prompt else None, seed=seed, ), ) if response_api.Resp is None: raise Exception(f"PixVerse request failed: '{response_api.ErrMsg}'") response_poll = await poll_op( cls, ApiEndpoint(path=f"/proxy/pixverse/video/result/{response_api.Resp.video_id}"), response_model=PixverseGenerationStatusResponse, completed_statuses=[PixverseStatus.successful], failed_statuses=[ PixverseStatus.contents_moderation, PixverseStatus.failed, PixverseStatus.deleted, ], status_extractor=lambda x: x.Resp.status, estimated_duration=AVERAGE_DURATION_T2V, ) return IO.NodeOutput(await download_url_to_video_output(response_poll.Resp.url)) PRICE_BADGE_VIDEO = IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["duration_seconds", "quality", "motion_mode"]), expr=""" ( $prices := { "5": { "1080p": {"normal": 1.2, "fast": 1.2}, "720p": {"normal": 0.6, "fast": 1.2}, "540p": {"normal": 0.45, "fast": 0.9}, "360p": {"normal": 0.45, "fast": 0.9} }, "8": { "1080p": {"normal": 1.2, "fast": 1.2}, "720p": {"normal": 1.2, "fast": 1.2}, "540p": {"normal": 0.9, "fast": 1.2}, "360p": {"normal": 0.9, "fast": 1.2} } }; $durPrices := $lookup($prices, $string(widgets.duration_seconds)); $qualityPrices := $lookup($durPrices, widgets.quality); $price := $lookup($qualityPrices, widgets.motion_mode); {"type":"usd","usd": $price ? $price : 0.9} ) """, ) class PixVerseExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[IO.ComfyNode]]: return [ PixverseTextToVideoNode, PixverseImageToVideoNode, PixverseTransitionVideoNode, PixverseTemplateNode, ] async def comfy_entrypoint() -> PixVerseExtension: return PixVerseExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api_nodes/nodes_pixverse.py", "license": "GNU General Public License v3.0", "lines": 429, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_api_nodes/nodes_recraft.py
from io import BytesIO import aiohttp import torch from PIL import UnidentifiedImageError from typing_extensions import override from comfy.utils import ProgressBar from comfy_api.latest import IO, ComfyExtension from comfy_api_nodes.apis.recraft import ( RECRAFT_V4_PRO_SIZES, RECRAFT_V4_SIZES, RecraftColor, RecraftColorChain, RecraftControls, RecraftCreateStyleRequest, RecraftCreateStyleResponse, RecraftImageGenerationRequest, RecraftImageGenerationResponse, RecraftImageSize, RecraftIO, RecraftStyle, RecraftStyleV3, get_v3_substyles, ) from comfy_api_nodes.util import ( ApiEndpoint, bytesio_to_image_tensor, download_url_as_bytesio, resize_mask_to_image, sync_op, tensor_to_bytesio, validate_string, ) from comfy_extras.nodes_images import SVG async def handle_recraft_file_request( cls: type[IO.ComfyNode], image: torch.Tensor, path: str, mask: torch.Tensor | None = None, total_pixels: int = 4096 * 4096, timeout: int = 1024, request=None, ) -> list[BytesIO]: """Handle sending common Recraft file-only request to get back file bytes.""" files = {"image": tensor_to_bytesio(image, total_pixels=total_pixels).read()} if mask is not None: files["mask"] = tensor_to_bytesio(mask, total_pixels=total_pixels).read() response = await sync_op( cls, endpoint=ApiEndpoint(path=path, method="POST"), response_model=RecraftImageGenerationResponse, data=request if request else None, files=files, content_type="multipart/form-data", multipart_parser=recraft_multipart_parser, max_retries=1, ) all_bytesio = [] if response.image is not None: all_bytesio.append(await download_url_as_bytesio(response.image.url, timeout=timeout)) else: for data in response.data: all_bytesio.append(await download_url_as_bytesio(data.url, timeout=timeout)) return all_bytesio def recraft_multipart_parser( data, parent_key=None, formatter: type[callable] | None = None, converted_to_check: list[list] | None = None, is_list: bool = False, return_mode: str = "formdata", # "dict" | "formdata" ) -> dict | aiohttp.FormData: """ Formats data such that multipart/form-data will work with aiohttp library when both files and data are present. The OpenAI client that Recraft uses has a bizarre way of serializing lists: It does NOT keep track of indeces of each list, so for background_color, that must be serialized as: 'background_color[rgb][]' = [0, 0, 255] where the array is assigned to a key that has '[]' at the end, to signal it's an array. This has the consequence of nested lists having the exact same key, forcing arrays to merge; all colors inputs fall under the same key: if 1 color -> 'controls[colors][][rgb][]' = [0, 0, 255] if 2 colors -> 'controls[colors][][rgb][]' = [0, 0, 255, 255, 0, 0] if 3 colors -> 'controls[colors][][rgb][]' = [0, 0, 255, 255, 0, 0, 0, 255, 0] etc. Whoever made this serialization up at OpenAI added the constraint that lists must be of uniform length on objects of same 'type'. """ # Modification of a function that handled a different type of multipart parsing, big ups: # https://gist.github.com/kazqvaizer/4cebebe5db654a414132809f9f88067b def handle_converted_lists(item, parent_key, lists_to_check=list[list]): # if list already exists, just extend list with data for check_list in lists_to_check: for conv_tuple in check_list: if conv_tuple[0] == parent_key and isinstance(conv_tuple[1], list): conv_tuple[1].append(formatter(item)) return True return False if converted_to_check is None: converted_to_check = [] effective_mode = return_mode if parent_key is None else "dict" if formatter is None: formatter = lambda v: v # Multipart representation of value if not isinstance(data, dict): # if list already exists, just extend list with data added = handle_converted_lists(data, parent_key, converted_to_check) if added: return {} # otherwise if is_list, create new list with data if is_list: return {parent_key: [formatter(data)]} # return new key with data return {parent_key: formatter(data)} converted = [] next_check = [converted] next_check.extend(converted_to_check) for key, value in data.items(): current_key = key if parent_key is None else f"{parent_key}[{key}]" if isinstance(value, dict): converted.extend(recraft_multipart_parser(value, current_key, formatter, next_check).items()) elif isinstance(value, list): for ind, list_value in enumerate(value): iter_key = f"{current_key}[]" converted.extend( recraft_multipart_parser(list_value, iter_key, formatter, next_check, is_list=True).items() ) else: converted.append((current_key, formatter(value))) if effective_mode == "formdata": fd = aiohttp.FormData() for k, v in dict(converted).items(): if isinstance(v, list): for item in v: fd.add_field(k, str(item)) else: fd.add_field(k, str(v)) return fd return dict(converted) class handle_recraft_image_output: """ Catch an exception related to receiving SVG data instead of image, when Infinite Style Library style_id is in use. """ def __init__(self): pass def __enter__(self): pass def __exit__(self, exc_type, exc_val, exc_tb): if exc_type is not None and exc_type is UnidentifiedImageError: raise Exception( "Received output data was not an image; likely an SVG. " "If you used style_id, make sure it is not a Vector art style." ) class RecraftColorRGBNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="RecraftColorRGB", display_name="Recraft Color RGB", category="api node/image/Recraft", description="Create Recraft Color by choosing specific RGB values.", inputs=[ IO.Int.Input("r", default=0, min=0, max=255, tooltip="Red value of color."), IO.Int.Input("g", default=0, min=0, max=255, tooltip="Green value of color."), IO.Int.Input("b", default=0, min=0, max=255, tooltip="Blue value of color."), IO.Custom(RecraftIO.COLOR).Input("recraft_color", optional=True), ], outputs=[ IO.Custom(RecraftIO.COLOR).Output(display_name="recraft_color"), ], ) @classmethod def execute(cls, r: int, g: int, b: int, recraft_color: RecraftColorChain = None) -> IO.NodeOutput: recraft_color = recraft_color.clone() if recraft_color else RecraftColorChain() recraft_color.add(RecraftColor(r, g, b)) return IO.NodeOutput(recraft_color) class RecraftControlsNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="RecraftControls", display_name="Recraft Controls", category="api node/image/Recraft", description="Create Recraft Controls for customizing Recraft generation.", inputs=[ IO.Custom(RecraftIO.COLOR).Input("colors", optional=True), IO.Custom(RecraftIO.COLOR).Input("background_color", optional=True), ], outputs=[ IO.Custom(RecraftIO.CONTROLS).Output(display_name="recraft_controls"), ], ) @classmethod def execute(cls, colors: RecraftColorChain = None, background_color: RecraftColorChain = None) -> IO.NodeOutput: return IO.NodeOutput(RecraftControls(colors=colors, background_color=background_color)) class RecraftStyleV3RealisticImageNode(IO.ComfyNode): RECRAFT_STYLE = RecraftStyleV3.realistic_image @classmethod def define_schema(cls): return IO.Schema( node_id="RecraftStyleV3RealisticImage", display_name="Recraft Style - Realistic Image", category="api node/image/Recraft", description="Select realistic_image style and optional substyle.", inputs=[ IO.Combo.Input("substyle", options=get_v3_substyles(cls.RECRAFT_STYLE)), ], outputs=[ IO.Custom(RecraftIO.STYLEV3).Output(display_name="recraft_style"), ], ) @classmethod def execute(cls, substyle: str) -> IO.NodeOutput: if substyle == "None": substyle = None return IO.NodeOutput(RecraftStyle(cls.RECRAFT_STYLE, substyle)) class RecraftStyleV3DigitalIllustrationNode(RecraftStyleV3RealisticImageNode): RECRAFT_STYLE = RecraftStyleV3.digital_illustration @classmethod def define_schema(cls): return IO.Schema( node_id="RecraftStyleV3DigitalIllustration", display_name="Recraft Style - Digital Illustration", category="api node/image/Recraft", description="Select realistic_image style and optional substyle.", inputs=[ IO.Combo.Input("substyle", options=get_v3_substyles(cls.RECRAFT_STYLE)), ], outputs=[ IO.Custom(RecraftIO.STYLEV3).Output(display_name="recraft_style"), ], ) class RecraftStyleV3VectorIllustrationNode(RecraftStyleV3RealisticImageNode): RECRAFT_STYLE = RecraftStyleV3.vector_illustration @classmethod def define_schema(cls): return IO.Schema( node_id="RecraftStyleV3VectorIllustrationNode", display_name="Recraft Style - Realistic Image", category="api node/image/Recraft", description="Select realistic_image style and optional substyle.", inputs=[ IO.Combo.Input("substyle", options=get_v3_substyles(cls.RECRAFT_STYLE)), ], outputs=[ IO.Custom(RecraftIO.STYLEV3).Output(display_name="recraft_style"), ], ) class RecraftStyleV3LogoRasterNode(RecraftStyleV3RealisticImageNode): RECRAFT_STYLE = RecraftStyleV3.logo_raster @classmethod def define_schema(cls): return IO.Schema( node_id="RecraftStyleV3LogoRaster", display_name="Recraft Style - Logo Raster", category="api node/image/Recraft", description="Select realistic_image style and optional substyle.", inputs=[ IO.Combo.Input("substyle", options=get_v3_substyles(cls.RECRAFT_STYLE, include_none=False)), ], outputs=[ IO.Custom(RecraftIO.STYLEV3).Output(display_name="recraft_style"), ], ) class RecraftStyleInfiniteStyleLibrary(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="RecraftStyleV3InfiniteStyleLibrary", display_name="Recraft Style - Infinite Style Library", category="api node/image/Recraft", description="Choose style based on preexisting UUID from Recraft's Infinite Style Library.", inputs=[ IO.String.Input("style_id", default="", tooltip="UUID of style from Infinite Style Library."), ], outputs=[ IO.Custom(RecraftIO.STYLEV3).Output(display_name="recraft_style"), ], ) @classmethod def execute(cls, style_id: str) -> IO.NodeOutput: if not style_id: raise Exception("The style_id input cannot be empty.") return IO.NodeOutput(RecraftStyle(style_id=style_id)) class RecraftCreateStyleNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="RecraftCreateStyleNode", display_name="Recraft Create Style", category="api node/image/Recraft", description="Create a custom style from reference images. " "Upload 1-5 images to use as style references. " "Total size of all images is limited to 5 MB.", inputs=[ IO.Combo.Input( "style", options=["realistic_image", "digital_illustration"], tooltip="The base style of the generated images.", ), IO.Autogrow.Input( "images", template=IO.Autogrow.TemplatePrefix( IO.Image.Input("image"), prefix="image", min=1, max=5, ), ), ], outputs=[ IO.String.Output(display_name="style_id"), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd": 0.04}""", ), ) @classmethod async def execute( cls, style: str, images: IO.Autogrow.Type, ) -> IO.NodeOutput: files = [] total_size = 0 max_total_size = 5 * 1024 * 1024 # 5 MB limit for i, img in enumerate(list(images.values())): file_bytes = tensor_to_bytesio(img, total_pixels=2048 * 2048, mime_type="image/webp").read() total_size += len(file_bytes) if total_size > max_total_size: raise Exception("Total size of all images exceeds 5 MB limit.") files.append((f"file{i + 1}", file_bytes)) response = await sync_op( cls, endpoint=ApiEndpoint(path="/proxy/recraft/styles", method="POST"), response_model=RecraftCreateStyleResponse, files=files, data=RecraftCreateStyleRequest(style=style), content_type="multipart/form-data", max_retries=1, ) return IO.NodeOutput(response.id) class RecraftTextToImageNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="RecraftTextToImageNode", display_name="Recraft Text to Image", category="api node/image/Recraft", description="Generates images synchronously based on prompt and resolution.", inputs=[ IO.String.Input("prompt", multiline=True, default="", tooltip="Prompt for the image generation."), IO.Combo.Input( "size", options=[res.value for res in RecraftImageSize], default=RecraftImageSize.res_1024x1024, tooltip="The size of the generated image.", ), IO.Int.Input( "n", default=1, min=1, max=6, tooltip="The number of images to generate.", ), IO.Int.Input( "seed", default=0, min=0, max=0xFFFFFFFFFFFFFFFF, control_after_generate=True, tooltip="Seed to determine if node should re-run; " "actual results are nondeterministic regardless of seed.", ), IO.Custom(RecraftIO.STYLEV3).Input("recraft_style", optional=True), IO.String.Input( "negative_prompt", default="", force_input=True, tooltip="An optional text description of undesired elements on an image.", optional=True, ), IO.Custom(RecraftIO.CONTROLS).Input( "recraft_controls", tooltip="Optional additional controls over the generation via the Recraft Controls node.", optional=True, ), ], outputs=[ IO.Image.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["n"]), expr="""{"type":"usd","usd": $round(0.04 * widgets.n, 2)}""", ), ) @classmethod async def execute( cls, prompt: str, size: str, n: int, seed, recraft_style: RecraftStyle = None, negative_prompt: str = None, recraft_controls: RecraftControls = None, ) -> IO.NodeOutput: validate_string(prompt, strip_whitespace=False, min_length=1, max_length=1000) default_style = RecraftStyle(RecraftStyleV3.realistic_image) if recraft_style is None: recraft_style = default_style controls_api = None if recraft_controls: controls_api = recraft_controls.create_api_model() if not negative_prompt: negative_prompt = None response = await sync_op( cls, ApiEndpoint(path="/proxy/recraft/image_generation", method="POST"), response_model=RecraftImageGenerationResponse, data=RecraftImageGenerationRequest( prompt=prompt, negative_prompt=negative_prompt, model="recraftv3", size=size, n=n, style=recraft_style.style, substyle=recraft_style.substyle, style_id=recraft_style.style_id, controls=controls_api, ), max_retries=1, ) images = [] for data in response.data: with handle_recraft_image_output(): image = bytesio_to_image_tensor(await download_url_as_bytesio(data.url, timeout=1024)) if len(image.shape) < 4: image = image.unsqueeze(0) images.append(image) return IO.NodeOutput(torch.cat(images, dim=0)) class RecraftImageToImageNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="RecraftImageToImageNode", display_name="Recraft Image to Image", category="api node/image/Recraft", description="Modify image based on prompt and strength.", inputs=[ IO.Image.Input("image"), IO.String.Input("prompt", multiline=True, default="", tooltip="Prompt for the image generation."), IO.Int.Input( "n", default=1, min=1, max=6, tooltip="The number of images to generate.", ), IO.Float.Input( "strength", default=0.5, min=0.0, max=1.0, step=0.01, tooltip="Defines the difference with the original image, should lie in [0, 1], " "where 0 means almost identical, and 1 means miserable similarity.", ), IO.Int.Input( "seed", default=0, min=0, max=0xFFFFFFFFFFFFFFFF, control_after_generate=True, tooltip="Seed to determine if node should re-run; " "actual results are nondeterministic regardless of seed.", ), IO.Custom(RecraftIO.STYLEV3).Input("recraft_style", optional=True), IO.String.Input( "negative_prompt", default="", force_input=True, tooltip="An optional text description of undesired elements on an image.", optional=True, ), IO.Custom(RecraftIO.CONTROLS).Input( "recraft_controls", tooltip="Optional additional controls over the generation via the Recraft Controls node.", optional=True, ), ], outputs=[ IO.Image.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["n"]), expr="""{"type":"usd","usd": $round(0.04 * widgets.n, 2)}""", ), ) @classmethod async def execute( cls, image: torch.Tensor, prompt: str, n: int, strength: float, seed, recraft_style: RecraftStyle = None, negative_prompt: str = None, recraft_controls: RecraftControls = None, ) -> IO.NodeOutput: validate_string(prompt, strip_whitespace=False, max_length=1000) default_style = RecraftStyle(RecraftStyleV3.realistic_image) if recraft_style is None: recraft_style = default_style controls_api = None if recraft_controls: controls_api = recraft_controls.create_api_model() if not negative_prompt: negative_prompt = None request = RecraftImageGenerationRequest( prompt=prompt, negative_prompt=negative_prompt, model="recraftv3", n=n, strength=round(strength, 2), style=recraft_style.style, substyle=recraft_style.substyle, style_id=recraft_style.style_id, controls=controls_api, ) images = [] total = image.shape[0] pbar = ProgressBar(total) for i in range(total): sub_bytes = await handle_recraft_file_request( cls, image=image[i], path="/proxy/recraft/images/imageToImage", request=request, ) with handle_recraft_image_output(): images.append(torch.cat([bytesio_to_image_tensor(x) for x in sub_bytes], dim=0)) pbar.update(1) return IO.NodeOutput(torch.cat(images, dim=0)) class RecraftImageInpaintingNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="RecraftImageInpaintingNode", display_name="Recraft Image Inpainting", category="api node/image/Recraft", description="Modify image based on prompt and mask.", inputs=[ IO.Image.Input("image"), IO.Mask.Input("mask"), IO.String.Input("prompt", multiline=True, default="", tooltip="Prompt for the image generation."), IO.Int.Input( "n", default=1, min=1, max=6, tooltip="The number of images to generate.", ), IO.Int.Input( "seed", default=0, min=0, max=0xFFFFFFFFFFFFFFFF, control_after_generate=True, tooltip="Seed to determine if node should re-run; " "actual results are nondeterministic regardless of seed.", ), IO.Custom(RecraftIO.STYLEV3).Input("recraft_style", optional=True), IO.String.Input( "negative_prompt", default="", force_input=True, tooltip="An optional text description of undesired elements on an image.", optional=True, ), ], outputs=[ IO.Image.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["n"]), expr="""{"type":"usd","usd": $round(0.04 * widgets.n, 2)}""", ), ) @classmethod async def execute( cls, image: torch.Tensor, mask: torch.Tensor, prompt: str, n: int, seed, recraft_style: RecraftStyle = None, negative_prompt: str = None, ) -> IO.NodeOutput: validate_string(prompt, strip_whitespace=False, max_length=1000) default_style = RecraftStyle(RecraftStyleV3.realistic_image) if recraft_style is None: recraft_style = default_style if not negative_prompt: negative_prompt = None request = RecraftImageGenerationRequest( prompt=prompt, negative_prompt=negative_prompt, model="recraftv3", n=n, style=recraft_style.style, substyle=recraft_style.substyle, style_id=recraft_style.style_id, ) # prepare mask tensor mask = resize_mask_to_image(mask, image, allow_gradient=False, add_channel_dim=True) images = [] total = image.shape[0] pbar = ProgressBar(total) for i in range(total): sub_bytes = await handle_recraft_file_request( cls, image=image[i], mask=mask[i : i + 1], path="/proxy/recraft/images/inpaint", request=request, ) with handle_recraft_image_output(): images.append(torch.cat([bytesio_to_image_tensor(x) for x in sub_bytes], dim=0)) pbar.update(1) return IO.NodeOutput(torch.cat(images, dim=0)) class RecraftTextToVectorNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="RecraftTextToVectorNode", display_name="Recraft Text to Vector", category="api node/image/Recraft", description="Generates SVG synchronously based on prompt and resolution.", inputs=[ IO.String.Input("prompt", default="", tooltip="Prompt for the image generation.", multiline=True), IO.Combo.Input("substyle", options=get_v3_substyles(RecraftStyleV3.vector_illustration)), IO.Combo.Input( "size", options=[res.value for res in RecraftImageSize], default=RecraftImageSize.res_1024x1024, tooltip="The size of the generated image.", ), IO.Int.Input("n", default=1, min=1, max=6, tooltip="The number of images to generate."), IO.Int.Input( "seed", default=0, min=0, max=0xFFFFFFFFFFFFFFFF, control_after_generate=True, tooltip="Seed to determine if node should re-run; " "actual results are nondeterministic regardless of seed.", ), IO.String.Input( "negative_prompt", default="", force_input=True, tooltip="An optional text description of undesired elements on an image.", optional=True, ), IO.Custom(RecraftIO.CONTROLS).Input( "recraft_controls", tooltip="Optional additional controls over the generation via the Recraft Controls node.", optional=True, ), ], outputs=[ IO.SVG.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["n"]), expr="""{"type":"usd","usd": $round(0.08 * widgets.n, 2)}""", ), ) @classmethod async def execute( cls, prompt: str, substyle: str, size: str, n: int, seed, negative_prompt: str = None, recraft_controls: RecraftControls = None, ) -> IO.NodeOutput: validate_string(prompt, strip_whitespace=False, max_length=1000) # create RecraftStyle so strings will be formatted properly (i.e. "None" will become None) recraft_style = RecraftStyle(RecraftStyleV3.vector_illustration, substyle=substyle) controls_api = None if recraft_controls: controls_api = recraft_controls.create_api_model() if not negative_prompt: negative_prompt = None response = await sync_op( cls, ApiEndpoint(path="/proxy/recraft/image_generation", method="POST"), response_model=RecraftImageGenerationResponse, data=RecraftImageGenerationRequest( prompt=prompt, negative_prompt=negative_prompt, model="recraftv3", size=size, n=n, style=recraft_style.style, substyle=recraft_style.substyle, controls=controls_api, ), max_retries=1, ) svg_data = [] for data in response.data: svg_data.append(await download_url_as_bytesio(data.url, timeout=1024)) return IO.NodeOutput(SVG(svg_data)) class RecraftVectorizeImageNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="RecraftVectorizeImageNode", display_name="Recraft Vectorize Image", category="api node/image/Recraft", description="Generates SVG synchronously from an input image.", inputs=[ IO.Image.Input("image"), ], outputs=[ IO.SVG.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(), expr="""{"type":"usd","usd": 0.01}""", ), ) @classmethod async def execute(cls, image: torch.Tensor) -> IO.NodeOutput: svgs = [] total = image.shape[0] pbar = ProgressBar(total) for i in range(total): sub_bytes = await handle_recraft_file_request( cls, image=image[i], path="/proxy/recraft/images/vectorize", ) svgs.append(SVG(sub_bytes)) pbar.update(1) return IO.NodeOutput(SVG.combine_all(svgs)) class RecraftReplaceBackgroundNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="RecraftReplaceBackgroundNode", display_name="Recraft Replace Background", category="api node/image/Recraft", description="Replace background on image, based on provided prompt.", inputs=[ IO.Image.Input("image"), IO.String.Input("prompt", tooltip="Prompt for the image generation.", default="", multiline=True), IO.Int.Input("n", default=1, min=1, max=6, tooltip="The number of images to generate."), IO.Int.Input( "seed", default=0, min=0, max=0xFFFFFFFFFFFFFFFF, control_after_generate=True, tooltip="Seed to determine if node should re-run; " "actual results are nondeterministic regardless of seed.", ), IO.Custom(RecraftIO.STYLEV3).Input("recraft_style", optional=True), IO.String.Input( "negative_prompt", default="", force_input=True, tooltip="An optional text description of undesired elements on an image.", optional=True, ), ], outputs=[ IO.Image.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.04}""", ), ) @classmethod async def execute( cls, image: torch.Tensor, prompt: str, n: int, seed, recraft_style: RecraftStyle = None, negative_prompt: str = None, ) -> IO.NodeOutput: default_style = RecraftStyle(RecraftStyleV3.realistic_image) if recraft_style is None: recraft_style = default_style if not negative_prompt: negative_prompt = None request = RecraftImageGenerationRequest( prompt=prompt, negative_prompt=negative_prompt, model="recraftv3", n=n, style=recraft_style.style, substyle=recraft_style.substyle, style_id=recraft_style.style_id, ) images = [] total = image.shape[0] pbar = ProgressBar(total) for i in range(total): sub_bytes = await handle_recraft_file_request( cls, image=image[i], path="/proxy/recraft/images/replaceBackground", request=request, ) images.append(torch.cat([bytesio_to_image_tensor(x) for x in sub_bytes], dim=0)) pbar.update(1) return IO.NodeOutput(torch.cat(images, dim=0)) class RecraftRemoveBackgroundNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="RecraftRemoveBackgroundNode", display_name="Recraft Remove Background", category="api node/image/Recraft", essentials_category="Image Tools", description="Remove background from image, and return processed image and mask.", inputs=[ IO.Image.Input("image"), ], outputs=[ IO.Image.Output(), IO.Mask.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.01}""", ), ) @classmethod async def execute(cls, image: torch.Tensor) -> IO.NodeOutput: images = [] total = image.shape[0] pbar = ProgressBar(total) for i in range(total): sub_bytes = await handle_recraft_file_request( cls, image=image[i], path="/proxy/recraft/images/removeBackground", ) images.append(torch.cat([bytesio_to_image_tensor(x) for x in sub_bytes], dim=0)) pbar.update(1) images_tensor = torch.cat(images, dim=0) # use alpha channel as masks, in B,H,W format masks_tensor = images_tensor[:, :, :, -1:].squeeze(-1) return IO.NodeOutput(images_tensor, masks_tensor) class RecraftCrispUpscaleNode(IO.ComfyNode): RECRAFT_PATH = "/proxy/recraft/images/crispUpscale" @classmethod def define_schema(cls): return IO.Schema( node_id="RecraftCrispUpscaleNode", display_name="Recraft Crisp Upscale Image", category="api node/image/Recraft", description="Upscale image synchronously.\n" "Enhances a given raster image using ‘crisp upscale’ tool, " "increasing image resolution, making the image sharper and cleaner.", inputs=[ IO.Image.Input("image"), ], outputs=[ IO.Image.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.004}""", ), ) @classmethod async def execute(cls, image: torch.Tensor) -> IO.NodeOutput: images = [] total = image.shape[0] pbar = ProgressBar(total) for i in range(total): sub_bytes = await handle_recraft_file_request( cls, image=image[i], path=cls.RECRAFT_PATH, ) images.append(torch.cat([bytesio_to_image_tensor(x) for x in sub_bytes], dim=0)) pbar.update(1) return IO.NodeOutput(torch.cat(images, dim=0)) class RecraftCreativeUpscaleNode(RecraftCrispUpscaleNode): RECRAFT_PATH = "/proxy/recraft/images/creativeUpscale" @classmethod def define_schema(cls): return IO.Schema( node_id="RecraftCreativeUpscaleNode", display_name="Recraft Creative Upscale Image", category="api node/image/Recraft", description="Upscale image synchronously.\n" "Enhances a given raster image using ‘creative upscale’ tool, " "boosting resolution with a focus on refining small details and faces.", inputs=[ IO.Image.Input("image"), ], outputs=[ IO.Image.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.25}""", ), ) class RecraftV4TextToImageNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="RecraftV4TextToImageNode", display_name="Recraft V4 Text to Image", category="api node/image/Recraft", description="Generates images using Recraft V4 or V4 Pro models.", inputs=[ IO.String.Input( "prompt", multiline=True, tooltip="Prompt for the image generation. Maximum 10,000 characters.", ), IO.String.Input( "negative_prompt", multiline=True, tooltip="An optional text description of undesired elements on an image.", ), IO.DynamicCombo.Input( "model", options=[ IO.DynamicCombo.Option( "recraftv4", [ IO.Combo.Input( "size", options=RECRAFT_V4_SIZES, default="1024x1024", tooltip="The size of the generated image.", ), ], ), IO.DynamicCombo.Option( "recraftv4_pro", [ IO.Combo.Input( "size", options=RECRAFT_V4_PRO_SIZES, default="2048x2048", tooltip="The size of the generated image.", ), ], ), ], tooltip="The model to use for generation.", ), IO.Int.Input( "n", default=1, min=1, max=6, tooltip="The number of images to generate.", ), IO.Int.Input( "seed", default=0, min=0, max=0xFFFFFFFFFFFFFFFF, control_after_generate=True, tooltip="Seed to determine if node should re-run; " "actual results are nondeterministic regardless of seed.", ), IO.Custom(RecraftIO.CONTROLS).Input( "recraft_controls", tooltip="Optional additional controls over the generation via the Recraft Controls node.", optional=True, ), ], outputs=[ IO.Image.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["model", "n"]), expr=""" ( $prices := {"recraftv4": 0.04, "recraftv4_pro": 0.25}; {"type":"usd","usd": $lookup($prices, widgets.model) * widgets.n} ) """, ), ) @classmethod async def execute( cls, prompt: str, negative_prompt: str, model: dict, n: int, seed: int, recraft_controls: RecraftControls | None = None, ) -> IO.NodeOutput: validate_string(prompt, strip_whitespace=False, min_length=1, max_length=10000) response = await sync_op( cls, ApiEndpoint(path="/proxy/recraft/image_generation", method="POST"), response_model=RecraftImageGenerationResponse, data=RecraftImageGenerationRequest( prompt=prompt, negative_prompt=negative_prompt if negative_prompt else None, model=model["model"], size=model["size"], n=n, controls=recraft_controls.create_api_model() if recraft_controls else None, ), max_retries=1, ) images = [] for data in response.data: with handle_recraft_image_output(): image = bytesio_to_image_tensor(await download_url_as_bytesio(data.url, timeout=1024)) if len(image.shape) < 4: image = image.unsqueeze(0) images.append(image) return IO.NodeOutput(torch.cat(images, dim=0)) class RecraftV4TextToVectorNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="RecraftV4TextToVectorNode", display_name="Recraft V4 Text to Vector", category="api node/image/Recraft", description="Generates SVG using Recraft V4 or V4 Pro models.", inputs=[ IO.String.Input( "prompt", multiline=True, tooltip="Prompt for the image generation. Maximum 10,000 characters.", ), IO.String.Input( "negative_prompt", multiline=True, tooltip="An optional text description of undesired elements on an image.", ), IO.DynamicCombo.Input( "model", options=[ IO.DynamicCombo.Option( "recraftv4", [ IO.Combo.Input( "size", options=RECRAFT_V4_SIZES, default="1024x1024", tooltip="The size of the generated image.", ), ], ), IO.DynamicCombo.Option( "recraftv4_pro", [ IO.Combo.Input( "size", options=RECRAFT_V4_PRO_SIZES, default="2048x2048", tooltip="The size of the generated image.", ), ], ), ], tooltip="The model to use for generation.", ), IO.Int.Input( "n", default=1, min=1, max=6, tooltip="The number of images to generate.", ), IO.Int.Input( "seed", default=0, min=0, max=0xFFFFFFFFFFFFFFFF, control_after_generate=True, tooltip="Seed to determine if node should re-run; " "actual results are nondeterministic regardless of seed.", ), IO.Custom(RecraftIO.CONTROLS).Input( "recraft_controls", tooltip="Optional additional controls over the generation via the Recraft Controls node.", optional=True, ), ], outputs=[ IO.SVG.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["model", "n"]), expr=""" ( $prices := {"recraftv4": 0.08, "recraftv4_pro": 0.30}; {"type":"usd","usd": $lookup($prices, widgets.model) * widgets.n} ) """, ), ) @classmethod async def execute( cls, prompt: str, negative_prompt: str, model: dict, n: int, seed: int, recraft_controls: RecraftControls | None = None, ) -> IO.NodeOutput: validate_string(prompt, strip_whitespace=False, min_length=1, max_length=10000) response = await sync_op( cls, ApiEndpoint(path="/proxy/recraft/image_generation", method="POST"), response_model=RecraftImageGenerationResponse, data=RecraftImageGenerationRequest( prompt=prompt, negative_prompt=negative_prompt if negative_prompt else None, model=model["model"], size=model["size"], n=n, style="vector_illustration", substyle=None, controls=recraft_controls.create_api_model() if recraft_controls else None, ), max_retries=1, ) svg_data = [] for data in response.data: svg_data.append(await download_url_as_bytesio(data.url, timeout=1024)) return IO.NodeOutput(SVG(svg_data)) class RecraftExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[IO.ComfyNode]]: return [ RecraftTextToImageNode, RecraftImageToImageNode, RecraftImageInpaintingNode, RecraftTextToVectorNode, RecraftVectorizeImageNode, RecraftRemoveBackgroundNode, RecraftReplaceBackgroundNode, RecraftCrispUpscaleNode, RecraftCreativeUpscaleNode, RecraftStyleV3RealisticImageNode, RecraftStyleV3DigitalIllustrationNode, RecraftStyleV3LogoRasterNode, RecraftStyleInfiniteStyleLibrary, RecraftCreateStyleNode, RecraftColorRGBNode, RecraftControlsNode, RecraftV4TextToImageNode, RecraftV4TextToVectorNode, ] async def comfy_entrypoint() -> RecraftExtension: return RecraftExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api_nodes/nodes_recraft.py", "license": "GNU General Public License v3.0", "lines": 1240, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_api_nodes/nodes_stability.py
from inspect import cleandoc from typing import Optional from typing_extensions import override from comfy_api.latest import ComfyExtension, Input, IO from comfy_api_nodes.apis.stability import ( StabilityUpscaleConservativeRequest, StabilityUpscaleCreativeRequest, StabilityAsyncResponse, StabilityResultsGetResponse, StabilityStable3_5Request, StabilityStableUltraRequest, StabilityStableUltraResponse, StabilityAspectRatio, Stability_SD3_5_Model, Stability_SD3_5_GenerationMode, get_stability_style_presets, StabilityTextToAudioRequest, StabilityAudioToAudioRequest, StabilityAudioInpaintRequest, StabilityAudioResponse, ) from comfy_api_nodes.util import ( validate_audio_duration, validate_string, audio_input_to_mp3, bytesio_to_image_tensor, tensor_to_bytesio, audio_bytes_to_audio_input, sync_op, poll_op, ApiEndpoint, ) import torch import base64 from io import BytesIO from enum import Enum class StabilityPollStatus(str, Enum): finished = "finished" in_progress = "in_progress" failed = "failed" def get_async_dummy_status(x: StabilityResultsGetResponse): if x.name is not None or x.errors is not None: return StabilityPollStatus.failed elif x.finish_reason is not None: return StabilityPollStatus.finished return StabilityPollStatus.in_progress class StabilityStableImageUltraNode(IO.ComfyNode): """ Generates images synchronously based on prompt and resolution. """ @classmethod def define_schema(cls): return IO.Schema( node_id="StabilityStableImageUltraNode", display_name="Stability AI Stable Image Ultra", category="api node/image/Stability AI", description=cleandoc(cls.__doc__ or ""), inputs=[ IO.String.Input( "prompt", multiline=True, default="", tooltip="What you wish to see in the output image. A strong, descriptive prompt that clearly defines" + "elements, colors, and subjects will lead to better results. " + "To control the weight of a given word use the format `(word:weight)`," + "where `word` is the word you'd like to control the weight of and `weight`" + "is a value between 0 and 1. For example: `The sky was a crisp (blue:0.3) and (green:0.8)`" + "would convey a sky that was blue and green, but more green than blue.", ), IO.Combo.Input( "aspect_ratio", options=StabilityAspectRatio, default=StabilityAspectRatio.ratio_1_1, tooltip="Aspect ratio of generated image.", ), IO.Combo.Input( "style_preset", options=get_stability_style_presets(), tooltip="Optional desired style of generated image.", advanced=True, ), IO.Int.Input( "seed", default=0, min=0, max=4294967294, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, tooltip="The random seed used for creating the noise.", ), IO.Image.Input( "image", optional=True, ), IO.String.Input( "negative_prompt", default="", tooltip="A blurb of text describing what you do not wish to see in the output image. This is an advanced feature.", force_input=True, optional=True, advanced=True, ), IO.Float.Input( "image_denoise", default=0.5, min=0.0, max=1.0, step=0.01, tooltip="Denoise of input image; 0.0 yields image identical to input, 1.0 is as if no image was provided at all.", optional=True, ), ], outputs=[ IO.Image.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.08}""", ), ) @classmethod async def execute( cls, prompt: str, aspect_ratio: str, style_preset: str, seed: int, image: Optional[torch.Tensor] = None, negative_prompt: str = "", image_denoise: Optional[float] = 0.5, ) -> IO.NodeOutput: validate_string(prompt, strip_whitespace=False) # prepare image binary if image present image_binary = None if image is not None: image_binary = tensor_to_bytesio(image, total_pixels=1504*1504).read() else: image_denoise = None if not negative_prompt: negative_prompt = None if style_preset == "None": style_preset = None files = { "image": image_binary } response_api = await sync_op( cls, ApiEndpoint(path="/proxy/stability/v2beta/stable-image/generate/ultra", method="POST"), response_model=StabilityStableUltraResponse, data=StabilityStableUltraRequest( prompt=prompt, negative_prompt=negative_prompt, aspect_ratio=aspect_ratio, seed=seed, strength=image_denoise, style_preset=style_preset, ), files=files, content_type="multipart/form-data", ) if response_api.finish_reason != "SUCCESS": raise Exception(f"Stable Image Ultra generation failed: {response_api.finish_reason}.") image_data = base64.b64decode(response_api.image) returned_image = bytesio_to_image_tensor(BytesIO(image_data)) return IO.NodeOutput(returned_image) class StabilityStableImageSD_3_5Node(IO.ComfyNode): """ Generates images synchronously based on prompt and resolution. """ @classmethod def define_schema(cls): return IO.Schema( node_id="StabilityStableImageSD_3_5Node", display_name="Stability AI Stable Diffusion 3.5 Image", category="api node/image/Stability AI", description=cleandoc(cls.__doc__ or ""), inputs=[ IO.String.Input( "prompt", multiline=True, default="", tooltip="What you wish to see in the output image. A strong, descriptive prompt that clearly defines elements, colors, and subjects will lead to better results.", ), IO.Combo.Input( "model", options=Stability_SD3_5_Model, ), IO.Combo.Input( "aspect_ratio", options=StabilityAspectRatio, default=StabilityAspectRatio.ratio_1_1, tooltip="Aspect ratio of generated image.", ), IO.Combo.Input( "style_preset", options=get_stability_style_presets(), tooltip="Optional desired style of generated image.", advanced=True, ), IO.Float.Input( "cfg_scale", default=4.0, min=1.0, max=10.0, step=0.1, tooltip="How strictly the diffusion process adheres to the prompt text (higher values keep your image closer to your prompt)", ), IO.Int.Input( "seed", default=0, min=0, max=4294967294, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, tooltip="The random seed used for creating the noise.", ), IO.Image.Input( "image", optional=True, ), IO.String.Input( "negative_prompt", default="", tooltip="Keywords of what you do not wish to see in the output image. This is an advanced feature.", force_input=True, optional=True, advanced=True, ), IO.Float.Input( "image_denoise", default=0.5, min=0.0, max=1.0, step=0.01, tooltip="Denoise of input image; 0.0 yields image identical to input, 1.0 is as if no image was provided at all.", optional=True, ), ], outputs=[ IO.Image.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["model"]), expr=""" ( $contains(widgets.model,"large") ? {"type":"usd","usd":0.065} : {"type":"usd","usd":0.035} ) """, ), ) @classmethod async def execute( cls, model: str, prompt: str, aspect_ratio: str, style_preset: str, seed: int, cfg_scale: float, image: Optional[torch.Tensor] = None, negative_prompt: str = "", image_denoise: Optional[float] = 0.5, ) -> IO.NodeOutput: validate_string(prompt, strip_whitespace=False) # prepare image binary if image present image_binary = None mode = Stability_SD3_5_GenerationMode.text_to_image if image is not None: image_binary = tensor_to_bytesio(image, total_pixels=1504*1504).read() mode = Stability_SD3_5_GenerationMode.image_to_image aspect_ratio = None else: image_denoise = None if not negative_prompt: negative_prompt = None if style_preset == "None": style_preset = None files = { "image": image_binary } response_api = await sync_op( cls, ApiEndpoint(path="/proxy/stability/v2beta/stable-image/generate/sd3", method="POST"), response_model=StabilityStableUltraResponse, data=StabilityStable3_5Request( prompt=prompt, negative_prompt=negative_prompt, aspect_ratio=aspect_ratio, seed=seed, strength=image_denoise, style_preset=style_preset, cfg_scale=cfg_scale, model=model, mode=mode, ), files=files, content_type="multipart/form-data", ) if response_api.finish_reason != "SUCCESS": raise Exception(f"Stable Diffusion 3.5 Image generation failed: {response_api.finish_reason}.") image_data = base64.b64decode(response_api.image) returned_image = bytesio_to_image_tensor(BytesIO(image_data)) return IO.NodeOutput(returned_image) class StabilityUpscaleConservativeNode(IO.ComfyNode): """ Upscale image with minimal alterations to 4K resolution. """ @classmethod def define_schema(cls): return IO.Schema( node_id="StabilityUpscaleConservativeNode", display_name="Stability AI Upscale Conservative", category="api node/image/Stability AI", description=cleandoc(cls.__doc__ or ""), inputs=[ IO.Image.Input("image"), IO.String.Input( "prompt", multiline=True, default="", tooltip="What you wish to see in the output image. A strong, descriptive prompt that clearly defines elements, colors, and subjects will lead to better results.", ), IO.Float.Input( "creativity", default=0.35, min=0.2, max=0.5, step=0.01, tooltip="Controls the likelihood of creating additional details not heavily conditioned by the init image.", ), IO.Int.Input( "seed", default=0, min=0, max=4294967294, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, tooltip="The random seed used for creating the noise.", ), IO.String.Input( "negative_prompt", default="", tooltip="Keywords of what you do not wish to see in the output image. This is an advanced feature.", force_input=True, optional=True, advanced=True, ), ], outputs=[ IO.Image.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.25}""", ), ) @classmethod async def execute( cls, image: torch.Tensor, prompt: str, creativity: float, seed: int, negative_prompt: str = "", ) -> IO.NodeOutput: validate_string(prompt, strip_whitespace=False) image_binary = tensor_to_bytesio(image, total_pixels=1024*1024).read() if not negative_prompt: negative_prompt = None files = { "image": image_binary } response_api = await sync_op( cls, ApiEndpoint(path="/proxy/stability/v2beta/stable-image/upscale/conservative", method="POST"), response_model=StabilityStableUltraResponse, data=StabilityUpscaleConservativeRequest( prompt=prompt, negative_prompt=negative_prompt, creativity=round(creativity,2), seed=seed, ), files=files, content_type="multipart/form-data", ) if response_api.finish_reason != "SUCCESS": raise Exception(f"Stability Upscale Conservative generation failed: {response_api.finish_reason}.") image_data = base64.b64decode(response_api.image) returned_image = bytesio_to_image_tensor(BytesIO(image_data)) return IO.NodeOutput(returned_image) class StabilityUpscaleCreativeNode(IO.ComfyNode): """ Upscale image with minimal alterations to 4K resolution. """ @classmethod def define_schema(cls): return IO.Schema( node_id="StabilityUpscaleCreativeNode", display_name="Stability AI Upscale Creative", category="api node/image/Stability AI", description=cleandoc(cls.__doc__ or ""), inputs=[ IO.Image.Input("image"), IO.String.Input( "prompt", multiline=True, default="", tooltip="What you wish to see in the output image. A strong, descriptive prompt that clearly defines elements, colors, and subjects will lead to better results.", ), IO.Float.Input( "creativity", default=0.3, min=0.1, max=0.5, step=0.01, tooltip="Controls the likelihood of creating additional details not heavily conditioned by the init image.", ), IO.Combo.Input( "style_preset", options=get_stability_style_presets(), tooltip="Optional desired style of generated image.", advanced=True, ), IO.Int.Input( "seed", default=0, min=0, max=4294967294, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, tooltip="The random seed used for creating the noise.", ), IO.String.Input( "negative_prompt", default="", tooltip="Keywords of what you do not wish to see in the output image. This is an advanced feature.", force_input=True, optional=True, advanced=True, ), ], outputs=[ IO.Image.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.25}""", ), ) @classmethod async def execute( cls, image: torch.Tensor, prompt: str, creativity: float, style_preset: str, seed: int, negative_prompt: str = "", ) -> IO.NodeOutput: validate_string(prompt, strip_whitespace=False) image_binary = tensor_to_bytesio(image, total_pixels=1024*1024).read() if not negative_prompt: negative_prompt = None if style_preset == "None": style_preset = None files = { "image": image_binary } response_api = await sync_op( cls, ApiEndpoint(path="/proxy/stability/v2beta/stable-image/upscale/creative", method="POST"), response_model=StabilityAsyncResponse, data=StabilityUpscaleCreativeRequest( prompt=prompt, negative_prompt=negative_prompt, creativity=round(creativity,2), style_preset=style_preset, seed=seed, ), files=files, content_type="multipart/form-data", ) response_poll = await poll_op( cls, ApiEndpoint(path=f"/proxy/stability/v2beta/results/{response_api.id}"), response_model=StabilityResultsGetResponse, poll_interval=3, status_extractor=lambda x: get_async_dummy_status(x), ) if response_poll.finish_reason != "SUCCESS": raise Exception(f"Stability Upscale Creative generation failed: {response_poll.finish_reason}.") image_data = base64.b64decode(response_poll.result) returned_image = bytesio_to_image_tensor(BytesIO(image_data)) return IO.NodeOutput(returned_image) class StabilityUpscaleFastNode(IO.ComfyNode): """ Quickly upscales an image via Stability API call to 4x its original size; intended for upscaling low-quality/compressed images. """ @classmethod def define_schema(cls): return IO.Schema( node_id="StabilityUpscaleFastNode", display_name="Stability AI Upscale Fast", category="api node/image/Stability AI", description=cleandoc(cls.__doc__ or ""), inputs=[ IO.Image.Input("image"), ], outputs=[ IO.Image.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.01}""", ), ) @classmethod async def execute(cls, image: torch.Tensor) -> IO.NodeOutput: image_binary = tensor_to_bytesio(image, total_pixels=4096*4096).read() files = { "image": image_binary } response_api = await sync_op( cls, ApiEndpoint(path="/proxy/stability/v2beta/stable-image/upscale/fast", method="POST"), response_model=StabilityStableUltraResponse, files=files, content_type="multipart/form-data", ) if response_api.finish_reason != "SUCCESS": raise Exception(f"Stability Upscale Fast failed: {response_api.finish_reason}.") image_data = base64.b64decode(response_api.image) returned_image = bytesio_to_image_tensor(BytesIO(image_data)) return IO.NodeOutput(returned_image) class StabilityTextToAudio(IO.ComfyNode): """Generates high-quality music and sound effects from text descriptions.""" @classmethod def define_schema(cls): return IO.Schema( node_id="StabilityTextToAudio", display_name="Stability AI Text To Audio", category="api node/audio/Stability AI", essentials_category="Audio", description=cleandoc(cls.__doc__ or ""), inputs=[ IO.Combo.Input( "model", options=["stable-audio-2.5"], ), IO.String.Input("prompt", multiline=True, default=""), IO.Int.Input( "duration", default=190, min=1, max=190, step=1, tooltip="Controls the duration in seconds of the generated audio.", optional=True, ), IO.Int.Input( "seed", default=0, min=0, max=4294967294, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, tooltip="The random seed used for generation.", optional=True, ), IO.Int.Input( "steps", default=8, min=4, max=8, step=1, tooltip="Controls the number of sampling steps.", optional=True, advanced=True, ), ], outputs=[ IO.Audio.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.2}""", ), ) @classmethod async def execute(cls, model: str, prompt: str, duration: int, seed: int, steps: int) -> IO.NodeOutput: validate_string(prompt, max_length=10000) payload = StabilityTextToAudioRequest(prompt=prompt, model=model, duration=duration, seed=seed, steps=steps) response_api = await sync_op( cls, ApiEndpoint(path="/proxy/stability/v2beta/audio/stable-audio-2/text-to-audio", method="POST"), response_model=StabilityAudioResponse, data=payload, content_type="multipart/form-data", ) if not response_api.audio: raise ValueError("No audio file was received in response.") return IO.NodeOutput(audio_bytes_to_audio_input(base64.b64decode(response_api.audio))) class StabilityAudioToAudio(IO.ComfyNode): """Transforms existing audio samples into new high-quality compositions using text instructions.""" @classmethod def define_schema(cls): return IO.Schema( node_id="StabilityAudioToAudio", display_name="Stability AI Audio To Audio", category="api node/audio/Stability AI", description=cleandoc(cls.__doc__ or ""), inputs=[ IO.Combo.Input( "model", options=["stable-audio-2.5"], ), IO.String.Input("prompt", multiline=True, default=""), IO.Audio.Input("audio", tooltip="Audio must be between 6 and 190 seconds long."), IO.Int.Input( "duration", default=190, min=1, max=190, step=1, tooltip="Controls the duration in seconds of the generated audio.", optional=True, ), IO.Int.Input( "seed", default=0, min=0, max=4294967294, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, tooltip="The random seed used for generation.", optional=True, ), IO.Int.Input( "steps", default=8, min=4, max=8, step=1, tooltip="Controls the number of sampling steps.", optional=True, advanced=True, ), IO.Float.Input( "strength", default=1, min=0.01, max=1.0, step=0.01, display_mode=IO.NumberDisplay.slider, tooltip="Parameter controls how much influence the audio parameter has on the generated audio.", optional=True, ), ], outputs=[ IO.Audio.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.2}""", ), ) @classmethod async def execute( cls, model: str, prompt: str, audio: Input.Audio, duration: int, seed: int, steps: int, strength: float ) -> IO.NodeOutput: validate_string(prompt, max_length=10000) validate_audio_duration(audio, 6, 190) payload = StabilityAudioToAudioRequest( prompt=prompt, model=model, duration=duration, seed=seed, steps=steps, strength=strength ) response_api = await sync_op( cls, ApiEndpoint(path="/proxy/stability/v2beta/audio/stable-audio-2/audio-to-audio", method="POST"), response_model=StabilityAudioResponse, data=payload, content_type="multipart/form-data", files={"audio": audio_input_to_mp3(audio)}, ) if not response_api.audio: raise ValueError("No audio file was received in response.") return IO.NodeOutput(audio_bytes_to_audio_input(base64.b64decode(response_api.audio))) class StabilityAudioInpaint(IO.ComfyNode): """Transforms part of existing audio sample using text instructions.""" @classmethod def define_schema(cls): return IO.Schema( node_id="StabilityAudioInpaint", display_name="Stability AI Audio Inpaint", category="api node/audio/Stability AI", description=cleandoc(cls.__doc__ or ""), inputs=[ IO.Combo.Input( "model", options=["stable-audio-2.5"], ), IO.String.Input("prompt", multiline=True, default=""), IO.Audio.Input("audio", tooltip="Audio must be between 6 and 190 seconds long."), IO.Int.Input( "duration", default=190, min=1, max=190, step=1, tooltip="Controls the duration in seconds of the generated audio.", optional=True, ), IO.Int.Input( "seed", default=0, min=0, max=4294967294, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, tooltip="The random seed used for generation.", optional=True, ), IO.Int.Input( "steps", default=8, min=4, max=8, step=1, tooltip="Controls the number of sampling steps.", optional=True, advanced=True, ), IO.Int.Input( "mask_start", default=30, min=0, max=190, step=1, optional=True, advanced=True, ), IO.Int.Input( "mask_end", default=190, min=0, max=190, step=1, optional=True, advanced=True, ), ], outputs=[ IO.Audio.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( expr="""{"type":"usd","usd":0.2}""", ), ) @classmethod async def execute( cls, model: str, prompt: str, audio: Input.Audio, duration: int, seed: int, steps: int, mask_start: int, mask_end: int, ) -> IO.NodeOutput: validate_string(prompt, max_length=10000) if mask_end <= mask_start: raise ValueError(f"Value of mask_end({mask_end}) should be greater then mask_start({mask_start})") validate_audio_duration(audio, 6, 190) payload = StabilityAudioInpaintRequest( prompt=prompt, model=model, duration=duration, seed=seed, steps=steps, mask_start=mask_start, mask_end=mask_end, ) response_api = await sync_op( cls, endpoint=ApiEndpoint(path="/proxy/stability/v2beta/audio/stable-audio-2/inpaint", method="POST"), response_model=StabilityAudioResponse, data=payload, content_type="multipart/form-data", files={"audio": audio_input_to_mp3(audio)}, ) if not response_api.audio: raise ValueError("No audio file was received in response.") return IO.NodeOutput(audio_bytes_to_audio_input(base64.b64decode(response_api.audio))) class StabilityExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[IO.ComfyNode]]: return [ StabilityStableImageUltraNode, StabilityStableImageSD_3_5Node, StabilityUpscaleConservativeNode, StabilityUpscaleCreativeNode, StabilityUpscaleFastNode, StabilityTextToAudio, StabilityAudioToAudio, StabilityAudioInpaint, ] async def comfy_entrypoint() -> StabilityExtension: return StabilityExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api_nodes/nodes_stability.py", "license": "GNU General Public License v3.0", "lines": 859, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:comfy_api_nodes/nodes_veo2.py
import base64 from io import BytesIO from typing_extensions import override from comfy_api.latest import IO, ComfyExtension, Input, InputImpl from comfy_api_nodes.apis.veo import ( VeoGenVidPollRequest, VeoGenVidPollResponse, VeoGenVidRequest, VeoGenVidResponse, VeoRequestInstance, VeoRequestInstanceImage, VeoRequestParameters, ) from comfy_api_nodes.util import ( ApiEndpoint, download_url_to_video_output, poll_op, sync_op, tensor_to_base64_string, ) AVERAGE_DURATION_VIDEO_GEN = 32 MODELS_MAP = { "veo-2.0-generate-001": "veo-2.0-generate-001", "veo-3.1-generate": "veo-3.1-generate-preview", "veo-3.1-fast-generate": "veo-3.1-fast-generate-preview", "veo-3.0-generate-001": "veo-3.0-generate-001", "veo-3.0-fast-generate-001": "veo-3.0-fast-generate-001", } class VeoVideoGenerationNode(IO.ComfyNode): """ Generates videos from text prompts using Google's Veo API. This node can create videos from text descriptions and optional image inputs, with control over parameters like aspect ratio, duration, and more. """ @classmethod def define_schema(cls): return IO.Schema( node_id="VeoVideoGenerationNode", display_name="Google Veo 2 Video Generation", category="api node/video/Veo", description="Generates videos from text prompts using Google's Veo 2 API", inputs=[ IO.String.Input( "prompt", multiline=True, default="", tooltip="Text description of the video", ), IO.Combo.Input( "aspect_ratio", options=["16:9", "9:16"], default="16:9", tooltip="Aspect ratio of the output video", ), IO.String.Input( "negative_prompt", multiline=True, default="", tooltip="Negative text prompt to guide what to avoid in the video", optional=True, ), IO.Int.Input( "duration_seconds", default=5, min=5, max=8, step=1, display_mode=IO.NumberDisplay.number, tooltip="Duration of the output video in seconds", optional=True, ), IO.Boolean.Input( "enhance_prompt", default=True, tooltip="Whether to enhance the prompt with AI assistance", optional=True, advanced=True, ), IO.Combo.Input( "person_generation", options=["ALLOW", "BLOCK"], default="ALLOW", tooltip="Whether to allow generating people in the video", optional=True, advanced=True, ), IO.Int.Input( "seed", default=0, min=0, max=0xFFFFFFFF, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, tooltip="Seed for video generation (0 for random)", optional=True, ), IO.Image.Input( "image", tooltip="Optional reference image to guide video generation", optional=True, ), IO.Combo.Input( "model", options=["veo-2.0-generate-001"], default="veo-2.0-generate-001", tooltip="Veo 2 model to use for video generation", optional=True, ), ], outputs=[ IO.Video.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["duration_seconds"]), expr="""{"type":"usd","usd": 0.5 * widgets.duration_seconds}""", ), ) @classmethod async def execute( cls, prompt, aspect_ratio="16:9", negative_prompt="", duration_seconds=5, enhance_prompt=True, person_generation="ALLOW", seed=0, image=None, model="veo-2.0-generate-001", generate_audio=False, ): model = MODELS_MAP[model] # Prepare the instances for the request instances = [] instance = {"prompt": prompt} # Add image if provided if image is not None: image_base64 = tensor_to_base64_string(image) if image_base64: instance["image"] = {"bytesBase64Encoded": image_base64, "mimeType": "image/png"} instances.append(instance) # Create parameters dictionary parameters = { "aspectRatio": aspect_ratio, "personGeneration": person_generation, "durationSeconds": duration_seconds, "enhancePrompt": enhance_prompt, } # Add optional parameters if provided if negative_prompt: parameters["negativePrompt"] = negative_prompt if seed > 0: parameters["seed"] = seed # Only add generateAudio for Veo 3 models if model.find("veo-2.0") == -1: parameters["generateAudio"] = generate_audio # force "enhance_prompt" to True for Veo3 models parameters["enhancePrompt"] = True initial_response = await sync_op( cls, ApiEndpoint(path=f"/proxy/veo/{model}/generate", method="POST"), response_model=VeoGenVidResponse, data=VeoGenVidRequest( instances=instances, parameters=parameters, ), ) def status_extractor(response): # Only return "completed" if the operation is done, regardless of success or failure # We'll check for errors after polling completes return "completed" if response.done else "pending" poll_response = await poll_op( cls, ApiEndpoint(path=f"/proxy/veo/{model}/poll", method="POST"), response_model=VeoGenVidPollResponse, status_extractor=status_extractor, data=VeoGenVidPollRequest( operationName=initial_response.name, ), poll_interval=5.0, estimated_duration=AVERAGE_DURATION_VIDEO_GEN, ) # Now check for errors in the final response # Check for error in poll response if poll_response.error: raise Exception(f"Veo API error: {poll_response.error.message} (code: {poll_response.error.code})") # Check for RAI filtered content if ( hasattr(poll_response.response, "raiMediaFilteredCount") and poll_response.response.raiMediaFilteredCount > 0 ): # Extract reason message if available if ( hasattr(poll_response.response, "raiMediaFilteredReasons") and poll_response.response.raiMediaFilteredReasons ): reason = poll_response.response.raiMediaFilteredReasons[0] error_message = f"Content filtered by Google's Responsible AI practices: {reason} ({poll_response.response.raiMediaFilteredCount} videos filtered.)" else: error_message = f"Content filtered by Google's Responsible AI practices ({poll_response.response.raiMediaFilteredCount} videos filtered.)" raise Exception(error_message) # Extract video data if ( poll_response.response and hasattr(poll_response.response, "videos") and poll_response.response.videos and len(poll_response.response.videos) > 0 ): video = poll_response.response.videos[0] # Check if video is provided as base64 or URL if hasattr(video, "bytesBase64Encoded") and video.bytesBase64Encoded: return IO.NodeOutput(InputImpl.VideoFromFile(BytesIO(base64.b64decode(video.bytesBase64Encoded)))) if hasattr(video, "gcsUri") and video.gcsUri: return IO.NodeOutput(await download_url_to_video_output(video.gcsUri)) raise Exception("Video returned but no data or URL was provided") raise Exception("Video generation completed but no video was returned") class Veo3VideoGenerationNode(VeoVideoGenerationNode): """ Generates videos from text prompts using Google's Veo 3 API. Supported models: - veo-3.0-generate-001 - veo-3.0-fast-generate-001 This node extends the base Veo node with Veo 3 specific features including audio generation and fixed 8-second duration. """ @classmethod def define_schema(cls): return IO.Schema( node_id="Veo3VideoGenerationNode", display_name="Google Veo 3 Video Generation", category="api node/video/Veo", description="Generates videos from text prompts using Google's Veo 3 API", inputs=[ IO.String.Input( "prompt", multiline=True, default="", tooltip="Text description of the video", ), IO.Combo.Input( "aspect_ratio", options=["16:9", "9:16"], default="16:9", tooltip="Aspect ratio of the output video", ), IO.String.Input( "negative_prompt", multiline=True, default="", tooltip="Negative text prompt to guide what to avoid in the video", optional=True, ), IO.Int.Input( "duration_seconds", default=8, min=8, max=8, step=1, display_mode=IO.NumberDisplay.number, tooltip="Duration of the output video in seconds (Veo 3 only supports 8 seconds)", optional=True, ), IO.Boolean.Input( "enhance_prompt", default=True, tooltip="This parameter is deprecated and ignored.", optional=True, advanced=True, ), IO.Combo.Input( "person_generation", options=["ALLOW", "BLOCK"], default="ALLOW", tooltip="Whether to allow generating people in the video", optional=True, advanced=True, ), IO.Int.Input( "seed", default=0, min=0, max=0xFFFFFFFF, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, tooltip="Seed for video generation (0 for random)", optional=True, ), IO.Image.Input( "image", tooltip="Optional reference image to guide video generation", optional=True, ), IO.Combo.Input( "model", options=[ "veo-3.1-generate", "veo-3.1-fast-generate", "veo-3.0-generate-001", "veo-3.0-fast-generate-001", ], default="veo-3.0-generate-001", tooltip="Veo 3 model to use for video generation", optional=True, ), IO.Boolean.Input( "generate_audio", default=False, tooltip="Generate audio for the video. Supported by all Veo 3 models.", optional=True, ), ], outputs=[ IO.Video.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["model", "generate_audio"]), expr=""" ( $m := widgets.model; $a := widgets.generate_audio; ($contains($m,"veo-3.0-fast-generate-001") or $contains($m,"veo-3.1-fast-generate")) ? {"type":"usd","usd": ($a ? 1.2 : 0.8)} : ($contains($m,"veo-3.0-generate-001") or $contains($m,"veo-3.1-generate")) ? {"type":"usd","usd": ($a ? 3.2 : 1.6)} : {"type":"range_usd","min_usd":0.8,"max_usd":3.2} ) """, ), ) class Veo3FirstLastFrameNode(IO.ComfyNode): @classmethod def define_schema(cls): return IO.Schema( node_id="Veo3FirstLastFrameNode", display_name="Google Veo 3 First-Last-Frame to Video", category="api node/video/Veo", description="Generate video using prompt and first and last frames.", inputs=[ IO.String.Input( "prompt", multiline=True, default="", tooltip="Text description of the video", ), IO.String.Input( "negative_prompt", multiline=True, default="", tooltip="Negative text prompt to guide what to avoid in the video", ), IO.Combo.Input("resolution", options=["720p", "1080p"]), IO.Combo.Input( "aspect_ratio", options=["16:9", "9:16"], default="16:9", tooltip="Aspect ratio of the output video", ), IO.Int.Input( "duration", default=8, min=4, max=8, step=2, display_mode=IO.NumberDisplay.slider, tooltip="Duration of the output video in seconds", ), IO.Int.Input( "seed", default=0, min=0, max=0xFFFFFFFF, step=1, display_mode=IO.NumberDisplay.number, control_after_generate=True, tooltip="Seed for video generation", ), IO.Image.Input("first_frame", tooltip="Start frame"), IO.Image.Input("last_frame", tooltip="End frame"), IO.Combo.Input( "model", options=["veo-3.1-generate", "veo-3.1-fast-generate"], default="veo-3.1-fast-generate", ), IO.Boolean.Input( "generate_audio", default=True, tooltip="Generate audio for the video.", ), ], outputs=[ IO.Video.Output(), ], hidden=[ IO.Hidden.auth_token_comfy_org, IO.Hidden.api_key_comfy_org, IO.Hidden.unique_id, ], is_api_node=True, price_badge=IO.PriceBadge( depends_on=IO.PriceBadgeDepends(widgets=["model", "generate_audio", "duration"]), expr=""" ( $prices := { "veo-3.1-fast-generate": { "audio": 0.15, "no_audio": 0.10 }, "veo-3.1-generate": { "audio": 0.40, "no_audio": 0.20 } }; $m := widgets.model; $ga := (widgets.generate_audio = "true"); $seconds := widgets.duration; $modelKey := $contains($m, "veo-3.1-fast-generate") ? "veo-3.1-fast-generate" : $contains($m, "veo-3.1-generate") ? "veo-3.1-generate" : ""; $audioKey := $ga ? "audio" : "no_audio"; $modelPrices := $lookup($prices, $modelKey); $pps := $lookup($modelPrices, $audioKey); ($pps != null) ? {"type":"usd","usd": $pps * $seconds} : {"type":"range_usd","min_usd": 0.4, "max_usd": 3.2} ) """, ), ) @classmethod async def execute( cls, prompt: str, negative_prompt: str, resolution: str, aspect_ratio: str, duration: int, seed: int, first_frame: Input.Image, last_frame: Input.Image, model: str, generate_audio: bool, ): model = MODELS_MAP[model] initial_response = await sync_op( cls, ApiEndpoint(path=f"/proxy/veo/{model}/generate", method="POST"), response_model=VeoGenVidResponse, data=VeoGenVidRequest( instances=[ VeoRequestInstance( prompt=prompt, image=VeoRequestInstanceImage( bytesBase64Encoded=tensor_to_base64_string(first_frame), mimeType="image/png" ), lastFrame=VeoRequestInstanceImage( bytesBase64Encoded=tensor_to_base64_string(last_frame), mimeType="image/png" ), ), ], parameters=VeoRequestParameters( aspectRatio=aspect_ratio, personGeneration="ALLOW", durationSeconds=duration, enhancePrompt=True, # cannot be False for Veo3 seed=seed, generateAudio=generate_audio, negativePrompt=negative_prompt, resolution=resolution, ), ), ) poll_response = await poll_op( cls, ApiEndpoint(path=f"/proxy/veo/{model}/poll", method="POST"), response_model=VeoGenVidPollResponse, status_extractor=lambda r: "completed" if r.done else "pending", data=VeoGenVidPollRequest( operationName=initial_response.name, ), poll_interval=5.0, estimated_duration=AVERAGE_DURATION_VIDEO_GEN, ) if poll_response.error: raise Exception(f"Veo API error: {poll_response.error.message} (code: {poll_response.error.code})") response = poll_response.response filtered_count = response.raiMediaFilteredCount if filtered_count: reasons = response.raiMediaFilteredReasons or [] reason_part = f": {reasons[0]}" if reasons else "" raise Exception( f"Content blocked by Google's Responsible AI filters{reason_part} " f"({filtered_count} video{'s' if filtered_count != 1 else ''} filtered)." ) if response.videos: video = response.videos[0] if video.bytesBase64Encoded: return IO.NodeOutput(InputImpl.VideoFromFile(BytesIO(base64.b64decode(video.bytesBase64Encoded)))) if video.gcsUri: return IO.NodeOutput(await download_url_to_video_output(video.gcsUri)) raise Exception("Video returned but no data or URL was provided") raise Exception("Video generation completed but no video was returned") class VeoExtension(ComfyExtension): @override async def get_node_list(self) -> list[type[IO.ComfyNode]]: return [ VeoVideoGenerationNode, Veo3VideoGenerationNode, Veo3FirstLastFrameNode, ] async def comfy_entrypoint() -> VeoExtension: return VeoExtension()
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_api_nodes/nodes_veo2.py", "license": "GNU General Public License v3.0", "lines": 520, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_complex
Comfy-Org/ComfyUI:tests-unit/comfy_api_test/input_impl_test.py
import io from comfy_api.input_impl.video_types import ( container_to_output_format, get_open_write_kwargs, ) from comfy_api.util import VideoContainer def test_container_to_output_format_empty_string(): """Test that an empty string input returns None. `None` arg allows default auto-detection.""" assert container_to_output_format("") is None def test_container_to_output_format_none(): """Test that None input returns None.""" assert container_to_output_format(None) is None def test_container_to_output_format_comma_separated(): """Test that a comma-separated list returns a valid singular format from the list.""" comma_separated_format = "mp4,mov,m4a" output_format = container_to_output_format(comma_separated_format) assert output_format in comma_separated_format def test_container_to_output_format_single(): """Test that a single format string (not comma-separated list) is returned as is.""" assert container_to_output_format("mp4") == "mp4" def test_get_open_write_kwargs_filepath_no_format(): """Test that 'format' kwarg is NOT set when dest is a file path.""" kwargs_auto = get_open_write_kwargs("output.mp4", "mp4", VideoContainer.AUTO) assert "format" not in kwargs_auto, "Format should not be set for file paths (AUTO)" kwargs_specific = get_open_write_kwargs("output.avi", "mp4", "avi") fail_msg = "Format should not be set for file paths (Specific)" assert "format" not in kwargs_specific, fail_msg def test_get_open_write_kwargs_base_options_mode(): """Test basic kwargs for file path: mode and movflags.""" kwargs = get_open_write_kwargs("output.mp4", "mp4", VideoContainer.AUTO) assert kwargs["mode"] == "w", "mode should be set to write" fail_msg = "movflags should be set to preserve custom metadata tags" assert "movflags" in kwargs["options"], fail_msg assert kwargs["options"]["movflags"] == "use_metadata_tags", fail_msg def test_get_open_write_kwargs_bytesio_auto_format(): """Test kwargs for BytesIO dest with AUTO format.""" dest = io.BytesIO() container_fmt = "mov,mp4,m4a" kwargs = get_open_write_kwargs(dest, container_fmt, VideoContainer.AUTO) assert kwargs["mode"] == "w" assert kwargs["options"]["movflags"] == "use_metadata_tags" fail_msg = ( "Format should be a valid format from the container's format list when AUTO" ) assert kwargs["format"] in container_fmt, fail_msg def test_get_open_write_kwargs_bytesio_specific_format(): """Test kwargs for BytesIO dest with a specific single format.""" dest = io.BytesIO() container_fmt = "avi" to_fmt = VideoContainer.MP4 kwargs = get_open_write_kwargs(dest, container_fmt, to_fmt) assert kwargs["mode"] == "w" assert kwargs["options"]["movflags"] == "use_metadata_tags" fail_msg = "Format should be the specified format (lowercased) when output format is not AUTO" assert kwargs["format"] == "mp4", fail_msg def test_get_open_write_kwargs_bytesio_specific_format_list(): """Test kwargs for BytesIO dest with a specific comma-separated format.""" dest = io.BytesIO() container_fmt = "avi" to_fmt = "mov,mp4,m4a" # A format string that is a list kwargs = get_open_write_kwargs(dest, container_fmt, to_fmt) assert kwargs["mode"] == "w" assert kwargs["options"]["movflags"] == "use_metadata_tags" fail_msg = "Format should be a valid format from the specified format list when output format is not AUTO" assert kwargs["format"] in to_fmt, fail_msg
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "tests-unit/comfy_api_test/input_impl_test.py", "license": "GNU General Public License v3.0", "lines": 65, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
test
Comfy-Org/ComfyUI:comfy_extras/nodes_preview_any.py
import json from comfy.comfy_types.node_typing import IO # Preview Any - original implement from # https://github.com/rgthree/rgthree-comfy/blob/main/py/display_any.py # upstream requested in https://github.com/Kosinkadink/rfcs/blob/main/rfcs/0000-corenodes.md#preview-nodes class PreviewAny(): @classmethod def INPUT_TYPES(cls): return { "required": {"source": (IO.ANY, {})}, } RETURN_TYPES = () FUNCTION = "main" OUTPUT_NODE = True CATEGORY = "utils" SEARCH_ALIASES = ["show output", "inspect", "debug", "print value", "show text"] def main(self, source=None): value = 'None' if isinstance(source, str): value = source elif isinstance(source, (int, float, bool)): value = str(source) elif source is not None: try: value = json.dumps(source, indent=4) except Exception: try: value = str(source) except Exception: value = 'source exists, but could not be serialized.' return {"ui": {"text": (value,)}} NODE_CLASS_MAPPINGS = { "PreviewAny": PreviewAny, } NODE_DISPLAY_NAME_MAPPINGS = { "PreviewAny": "Preview as Text", }
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy_extras/nodes_preview_any.py", "license": "GNU General Public License v3.0", "lines": 37, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple
Comfy-Org/ComfyUI:comfy/ldm/chroma/layers.py
import torch from torch import Tensor, nn from comfy.ldm.flux.layers import ( MLPEmbedder, ModulationOut, ) # TODO: remove this in a few months SingleStreamBlock = None DoubleStreamBlock = None class ChromaModulationOut(ModulationOut): @classmethod def from_offset(cls, tensor: torch.Tensor, offset: int = 0) -> ModulationOut: return cls( shift=tensor[:, offset : offset + 1, :], scale=tensor[:, offset + 1 : offset + 2, :], gate=tensor[:, offset + 2 : offset + 3, :], ) class Approximator(nn.Module): def __init__(self, in_dim: int, out_dim: int, hidden_dim: int, n_layers = 5, dtype=None, device=None, operations=None): super().__init__() self.in_proj = operations.Linear(in_dim, hidden_dim, bias=True, dtype=dtype, device=device) self.layers = nn.ModuleList([MLPEmbedder(hidden_dim, hidden_dim, dtype=dtype, device=device, operations=operations) for x in range( n_layers)]) self.norms = nn.ModuleList([operations.RMSNorm(hidden_dim, dtype=dtype, device=device) for x in range( n_layers)]) self.out_proj = operations.Linear(hidden_dim, out_dim, dtype=dtype, device=device) @property def device(self): # Get the device of the module (assumes all parameters are on the same device) return next(self.parameters()).device def forward(self, x: Tensor) -> Tensor: x = self.in_proj(x) for layer, norms in zip(self.layers, self.norms): x = x + layer(norms(x)) x = self.out_proj(x) return x class LastLayer(nn.Module): def __init__(self, hidden_size: int, patch_size: int, out_channels: int, dtype=None, device=None, operations=None): super().__init__() self.norm_final = operations.LayerNorm(hidden_size, elementwise_affine=False, eps=1e-6, dtype=dtype, device=device) self.linear = operations.Linear(hidden_size, out_channels, bias=True, dtype=dtype, device=device) def forward(self, x: Tensor, vec: Tensor) -> Tensor: shift, scale = vec shift = shift.squeeze(1) scale = scale.squeeze(1) x = torch.addcmul(shift[:, None, :], 1 + scale[:, None, :], self.norm_final(x)) x = self.linear(x) return x
{ "repo_id": "Comfy-Org/ComfyUI", "file_path": "comfy/ldm/chroma/layers.py", "license": "GNU General Public License v3.0", "lines": 46, "canary_id": -1, "canary_value": "", "pii_type": "", "provider": "", "regex_pattern": "", "repetition": -1, "template": "" }
function_simple