| import io
|
| import logging
|
| import base64
|
| import requests
|
| import torch
|
| from typing import Optional
|
|
|
| from comfy.comfy_types.node_typing import IO, ComfyNodeABC
|
| from comfy_api.input_impl.video_types import VideoFromFile
|
| from comfy_api_nodes.apis import (
|
| Veo2GenVidRequest,
|
| Veo2GenVidResponse,
|
| Veo2GenVidPollRequest,
|
| Veo2GenVidPollResponse
|
| )
|
| from comfy_api_nodes.apis.client import (
|
| ApiEndpoint,
|
| HttpMethod,
|
| SynchronousOperation,
|
| PollingOperation,
|
| )
|
|
|
| from comfy_api_nodes.apinode_utils import (
|
| downscale_image_tensor,
|
| tensor_to_base64_string
|
| )
|
|
|
| AVERAGE_DURATION_VIDEO_GEN = 32
|
|
|
| def convert_image_to_base64(image: torch.Tensor):
|
| if image is None:
|
| return None
|
|
|
| scaled_image = downscale_image_tensor(image, total_pixels=2048*2048)
|
| return tensor_to_base64_string(scaled_image)
|
|
|
|
|
| def get_video_url_from_response(poll_response: Veo2GenVidPollResponse) -> Optional[str]:
|
| if (
|
| poll_response.response
|
| and hasattr(poll_response.response, "videos")
|
| and poll_response.response.videos
|
| and len(poll_response.response.videos) > 0
|
| ):
|
| video = poll_response.response.videos[0]
|
| else:
|
| return None
|
| if hasattr(video, "gcsUri") and video.gcsUri:
|
| return str(video.gcsUri)
|
| return None
|
|
|
|
|
| class VeoVideoGenerationNode(ComfyNodeABC):
|
| """
|
| Generates videos from text prompts using Google's Veo API.
|
|
|
| This node can create videos from text descriptions and optional image inputs,
|
| with control over parameters like aspect ratio, duration, and more.
|
| """
|
|
|
| @classmethod
|
| def INPUT_TYPES(s):
|
| return {
|
| "required": {
|
| "prompt": (
|
| IO.STRING,
|
| {
|
| "multiline": True,
|
| "default": "",
|
| "tooltip": "Text description of the video",
|
| },
|
| ),
|
| "aspect_ratio": (
|
| IO.COMBO,
|
| {
|
| "options": ["16:9", "9:16"],
|
| "default": "16:9",
|
| "tooltip": "Aspect ratio of the output video",
|
| },
|
| ),
|
| },
|
| "optional": {
|
| "negative_prompt": (
|
| IO.STRING,
|
| {
|
| "multiline": True,
|
| "default": "",
|
| "tooltip": "Negative text prompt to guide what to avoid in the video",
|
| },
|
| ),
|
| "duration_seconds": (
|
| IO.INT,
|
| {
|
| "default": 5,
|
| "min": 5,
|
| "max": 8,
|
| "step": 1,
|
| "display": "number",
|
| "tooltip": "Duration of the output video in seconds",
|
| },
|
| ),
|
| "enhance_prompt": (
|
| IO.BOOLEAN,
|
| {
|
| "default": True,
|
| "tooltip": "Whether to enhance the prompt with AI assistance",
|
| }
|
| ),
|
| "person_generation": (
|
| IO.COMBO,
|
| {
|
| "options": ["ALLOW", "BLOCK"],
|
| "default": "ALLOW",
|
| "tooltip": "Whether to allow generating people in the video",
|
| },
|
| ),
|
| "seed": (
|
| IO.INT,
|
| {
|
| "default": 0,
|
| "min": 0,
|
| "max": 0xFFFFFFFF,
|
| "step": 1,
|
| "display": "number",
|
| "control_after_generate": True,
|
| "tooltip": "Seed for video generation (0 for random)",
|
| },
|
| ),
|
| "image": (IO.IMAGE, {
|
| "default": None,
|
| "tooltip": "Optional reference image to guide video generation",
|
| }),
|
| },
|
| "hidden": {
|
| "auth_token": "AUTH_TOKEN_COMFY_ORG",
|
| "comfy_api_key": "API_KEY_COMFY_ORG",
|
| "unique_id": "UNIQUE_ID",
|
| },
|
| }
|
|
|
| RETURN_TYPES = (IO.VIDEO,)
|
| FUNCTION = "generate_video"
|
| CATEGORY = "api node/video/Veo"
|
| DESCRIPTION = "Generates videos from text prompts using Google's Veo API"
|
| API_NODE = True
|
|
|
| def generate_video(
|
| self,
|
| prompt,
|
| aspect_ratio="16:9",
|
| negative_prompt="",
|
| duration_seconds=5,
|
| enhance_prompt=True,
|
| person_generation="ALLOW",
|
| seed=0,
|
| image=None,
|
| unique_id: Optional[str] = None,
|
| **kwargs,
|
| ):
|
|
|
| instances = []
|
|
|
| instance = {
|
| "prompt": prompt
|
| }
|
|
|
|
|
| if image is not None:
|
| image_base64 = convert_image_to_base64(image)
|
| if image_base64:
|
| instance["image"] = {
|
| "bytesBase64Encoded": image_base64,
|
| "mimeType": "image/png"
|
| }
|
|
|
| instances.append(instance)
|
|
|
|
|
| parameters = {
|
| "aspectRatio": aspect_ratio,
|
| "personGeneration": person_generation,
|
| "durationSeconds": duration_seconds,
|
| "enhancePrompt": enhance_prompt,
|
| }
|
|
|
|
|
| if negative_prompt:
|
| parameters["negativePrompt"] = negative_prompt
|
| if seed > 0:
|
| parameters["seed"] = seed
|
|
|
|
|
| initial_operation = SynchronousOperation(
|
| endpoint=ApiEndpoint(
|
| path="/proxy/veo/generate",
|
| method=HttpMethod.POST,
|
| request_model=Veo2GenVidRequest,
|
| response_model=Veo2GenVidResponse
|
| ),
|
| request=Veo2GenVidRequest(
|
| instances=instances,
|
| parameters=parameters
|
| ),
|
| auth_kwargs=kwargs,
|
| )
|
|
|
| initial_response = initial_operation.execute()
|
| operation_name = initial_response.name
|
|
|
| logging.info(f"Veo generation started with operation name: {operation_name}")
|
|
|
|
|
| def status_extractor(response):
|
|
|
|
|
| return "completed" if response.done else "pending"
|
|
|
|
|
| def progress_extractor(response):
|
|
|
| return None
|
|
|
|
|
| poll_operation = PollingOperation(
|
| poll_endpoint=ApiEndpoint(
|
| path="/proxy/veo/poll",
|
| method=HttpMethod.POST,
|
| request_model=Veo2GenVidPollRequest,
|
| response_model=Veo2GenVidPollResponse
|
| ),
|
| completed_statuses=["completed"],
|
| failed_statuses=[],
|
| status_extractor=status_extractor,
|
| progress_extractor=progress_extractor,
|
| request=Veo2GenVidPollRequest(
|
| operationName=operation_name
|
| ),
|
| auth_kwargs=kwargs,
|
| poll_interval=5.0,
|
| result_url_extractor=get_video_url_from_response,
|
| node_id=unique_id,
|
| estimated_duration=AVERAGE_DURATION_VIDEO_GEN,
|
| )
|
|
|
|
|
| poll_response = poll_operation.execute()
|
|
|
|
|
|
|
| if hasattr(poll_response, 'error') and poll_response.error:
|
| error_message = f"Veo API error: {poll_response.error.message} (code: {poll_response.error.code})"
|
| logging.error(error_message)
|
| raise Exception(error_message)
|
|
|
|
|
| if (hasattr(poll_response.response, 'raiMediaFilteredCount') and
|
| poll_response.response.raiMediaFilteredCount > 0):
|
|
|
|
|
| if (hasattr(poll_response.response, 'raiMediaFilteredReasons') and
|
| poll_response.response.raiMediaFilteredReasons):
|
| reason = poll_response.response.raiMediaFilteredReasons[0]
|
| error_message = f"Content filtered by Google's Responsible AI practices: {reason} ({poll_response.response.raiMediaFilteredCount} videos filtered.)"
|
| else:
|
| error_message = f"Content filtered by Google's Responsible AI practices ({poll_response.response.raiMediaFilteredCount} videos filtered.)"
|
|
|
| logging.error(error_message)
|
| raise Exception(error_message)
|
|
|
|
|
| video_data = None
|
| if poll_response.response and hasattr(poll_response.response, 'videos') and poll_response.response.videos and len(poll_response.response.videos) > 0:
|
| video = poll_response.response.videos[0]
|
|
|
|
|
| if hasattr(video, 'bytesBase64Encoded') and video.bytesBase64Encoded:
|
|
|
| video_data = base64.b64decode(video.bytesBase64Encoded)
|
| elif hasattr(video, 'gcsUri') and video.gcsUri:
|
|
|
| video_url = video.gcsUri
|
| video_response = requests.get(video_url)
|
| video_data = video_response.content
|
| else:
|
| raise Exception("Video returned but no data or URL was provided")
|
| else:
|
| raise Exception("Video generation completed but no video was returned")
|
|
|
| if not video_data:
|
| raise Exception("No video data was returned")
|
|
|
| logging.info("Video generation completed successfully")
|
|
|
|
|
| video_io = io.BytesIO(video_data)
|
|
|
|
|
| return (VideoFromFile(video_io),)
|
|
|
|
|
|
|
| NODE_CLASS_MAPPINGS = {
|
| "VeoVideoGenerationNode": VeoVideoGenerationNode,
|
| }
|
|
|
| NODE_DISPLAY_NAME_MAPPINGS = {
|
| "VeoVideoGenerationNode": "Google Veo2 Video Generation",
|
| }
|
|
|