{"repo_name": "Wan2.1", "file_name": "/Wan2.1/gradio/vace.py", "inference_info": {"prefix_code": "# -*- coding: utf-8 -*-\n# Copyright (c) Alibaba, Inc. and its affiliates.\n\nimport argparse\nimport datetime\nimport os\nimport sys\n\nimport imageio\nimport numpy as np\nimport torch\n\nimport gradio as gr\n\nsys.path.insert(\n 0, os.path.sep.join(os.path.realpath(__file__).split(os.path.sep)[:-2]))\nimport wan\nfrom wan import WanVace, WanVaceMP\nfrom wan.configs import SIZE_CONFIGS, WAN_CONFIGS\n\n\nclass FixedSizeQueue:\n\n def __init__(self, max_size):\n self.max_size = max_size\n self.queue = []\n\n def add(self, item):\n self.queue.insert(0, item)\n if len(self.queue) > self.max_size:\n self.queue.pop()\n\n def get(self):\n return self.queue\n\n def __repr__(self):\n return str(self.queue)\n\n\n", "suffix_code": "\n\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser(\n description='Argparser for VACE-WAN Demo:\\n')\n parser.add_argument(\n '--server_port', dest='server_port', help='', type=int, default=7860)\n parser.add_argument(\n '--server_name', dest='server_name', help='', default='0.0.0.0')\n parser.add_argument('--root_path', dest='root_path', help='', default=None)\n parser.add_argument('--save_dir', dest='save_dir', help='', default='cache')\n parser.add_argument(\n \"--mp\",\n action=\"store_true\",\n help=\"Use Multi-GPUs\",\n )\n parser.add_argument(\n \"--model_name\",\n type=str,\n default=\"vace-14B\",\n choices=list(WAN_CONFIGS.keys()),\n help=\"The model name to run.\")\n parser.add_argument(\n \"--ulysses_size\",\n type=int,\n default=1,\n help=\"The size of the ulysses parallelism in DiT.\")\n parser.add_argument(\n \"--ring_size\",\n type=int,\n default=1,\n help=\"The size of the ring attention parallelism in DiT.\")\n parser.add_argument(\n \"--ckpt_dir\",\n type=str,\n # default='models/VACE-Wan2.1-1.3B-Preview',\n default='models/Wan2.1-VACE-14B/',\n help=\"The path to the checkpoint directory.\",\n )\n parser.add_argument(\n \"--offload_to_cpu\",\n action=\"store_true\",\n help=\"Offloading unnecessary computations to CPU.\",\n )\n\n args = parser.parse_args()\n\n if not os.path.exists(args.save_dir):\n os.makedirs(args.save_dir, exist_ok=True)\n\n with gr.Blocks() as demo:\n infer_gr = VACEInference(\n args, skip_load=False, gallery_share=True, gallery_share_limit=5)\n infer_gr.create_ui()\n infer_gr.set_callbacks()\n allowed_paths = [args.save_dir]\n demo.queue(status_update_rate=1).launch(\n server_name=args.server_name,\n server_port=args.server_port,\n root_path=args.root_path,\n allowed_paths=allowed_paths,\n show_error=True,\n debug=True)\n", "middle_code": "class VACEInference:\n def __init__(self,\n cfg,\n skip_load=False,\n gallery_share=True,\n gallery_share_limit=5):\n self.cfg = cfg\n self.save_dir = cfg.save_dir\n self.gallery_share = gallery_share\n self.gallery_share_data = FixedSizeQueue(max_size=gallery_share_limit)\n if not skip_load:\n if not args.mp:\n self.pipe = WanVace(\n config=WAN_CONFIGS[cfg.model_name],\n checkpoint_dir=cfg.ckpt_dir,\n device_id=0,\n rank=0,\n t5_fsdp=False,\n dit_fsdp=False,\n use_usp=False,\n )\n else:\n self.pipe = WanVaceMP(\n config=WAN_CONFIGS[cfg.model_name],\n checkpoint_dir=cfg.ckpt_dir,\n use_usp=True,\n ulysses_size=cfg.ulysses_size,\n ring_size=cfg.ring_size)\n def create_ui(self, *args, **kwargs):\n gr.Markdown()\n with gr.Row(variant='panel', equal_height=True):\n with gr.Column(scale=1, min_width=0):\n self.src_video = gr.Video(\n label=\"src_video\",\n sources=['upload'],\n value=None,\n interactive=True)\n with gr.Column(scale=1, min_width=0):\n self.src_mask = gr.Video(\n label=\"src_mask\",\n sources=['upload'],\n value=None,\n interactive=True)\n with gr.Row(variant='panel', equal_height=True):\n with gr.Column(scale=1, min_width=0):\n with gr.Row(equal_height=True):\n self.src_ref_image_1 = gr.Image(\n label='src_ref_image_1',\n height=200,\n interactive=True,\n type='filepath',\n image_mode='RGB',\n sources=['upload'],\n elem_id=\"src_ref_image_1\",\n format='png')\n self.src_ref_image_2 = gr.Image(\n label='src_ref_image_2',\n height=200,\n interactive=True,\n type='filepath',\n image_mode='RGB',\n sources=['upload'],\n elem_id=\"src_ref_image_2\",\n format='png')\n self.src_ref_image_3 = gr.Image(\n label='src_ref_image_3',\n height=200,\n interactive=True,\n type='filepath',\n image_mode='RGB',\n sources=['upload'],\n elem_id=\"src_ref_image_3\",\n format='png')\n with gr.Row(variant='panel', equal_height=True):\n with gr.Column(scale=1):\n self.prompt = gr.Textbox(\n show_label=False,\n placeholder=\"positive_prompt_input\",\n elem_id='positive_prompt',\n container=True,\n autofocus=True,\n elem_classes='type_row',\n visible=True,\n lines=2)\n self.negative_prompt = gr.Textbox(\n show_label=False,\n value=self.pipe.config.sample_neg_prompt,\n placeholder=\"negative_prompt_input\",\n elem_id='negative_prompt',\n container=True,\n autofocus=False,\n elem_classes='type_row',\n visible=True,\n interactive=True,\n lines=1)\n with gr.Row(variant='panel', equal_height=True):\n with gr.Column(scale=1, min_width=0):\n with gr.Row(equal_height=True):\n self.shift_scale = gr.Slider(\n label='shift_scale',\n minimum=0.0,\n maximum=100.0,\n step=1.0,\n value=16.0,\n interactive=True)\n self.sample_steps = gr.Slider(\n label='sample_steps',\n minimum=1,\n maximum=100,\n step=1,\n value=25,\n interactive=True)\n self.context_scale = gr.Slider(\n label='context_scale',\n minimum=0.0,\n maximum=2.0,\n step=0.1,\n value=1.0,\n interactive=True)\n self.guide_scale = gr.Slider(\n label='guide_scale',\n minimum=1,\n maximum=10,\n step=0.5,\n value=5.0,\n interactive=True)\n self.infer_seed = gr.Slider(\n minimum=-1, maximum=10000000, value=2025, label=\"Seed\")\n with gr.Accordion(label=\"Usable without source video\", open=False):\n with gr.Row(equal_height=True):\n self.output_height = gr.Textbox(\n label='resolutions_height',\n value=720,\n interactive=True)\n self.output_width = gr.Textbox(\n label='resolutions_width',\n value=1280,\n interactive=True)\n self.frame_rate = gr.Textbox(\n label='frame_rate', value=16, interactive=True)\n self.num_frames = gr.Textbox(\n label='num_frames', value=81, interactive=True)\n with gr.Row(equal_height=True):\n with gr.Column(scale=5):\n self.generate_button = gr.Button(\n value='Run',\n elem_classes='type_row',\n elem_id='generate_button',\n visible=True)\n with gr.Column(scale=1):\n self.refresh_button = gr.Button(value='\\U0001f504') \n self.output_gallery = gr.Gallery(\n label=\"output_gallery\",\n value=[],\n interactive=False,\n allow_preview=True,\n preview=True)\n def generate(self, output_gallery, src_video, src_mask, src_ref_image_1,\n src_ref_image_2, src_ref_image_3, prompt, negative_prompt,\n shift_scale, sample_steps, context_scale, guide_scale,\n infer_seed, output_height, output_width, frame_rate,\n num_frames):\n output_height, output_width, frame_rate, num_frames = int(\n output_height), int(output_width), int(frame_rate), int(num_frames)\n src_ref_images = [\n x for x in [src_ref_image_1, src_ref_image_2, src_ref_image_3]\n if x is not None\n ]\n src_video, src_mask, src_ref_images = self.pipe.prepare_source(\n [src_video], [src_mask], [src_ref_images],\n num_frames=num_frames,\n image_size=SIZE_CONFIGS[f\"{output_width}*{output_height}\"],\n device=self.pipe.device)\n video = self.pipe.generate(\n prompt,\n src_video,\n src_mask,\n src_ref_images,\n size=(output_width, output_height),\n context_scale=context_scale,\n shift=shift_scale,\n sampling_steps=sample_steps,\n guide_scale=guide_scale,\n n_prompt=negative_prompt,\n seed=infer_seed,\n offload_model=True)\n name = '{0:%Y%m%d%-H%M%S}'.format(datetime.datetime.now())\n video_path = os.path.join(self.save_dir, f'cur_gallery_{name}.mp4')\n video_frames = (\n torch.clamp(video / 2 + 0.5, min=0.0, max=1.0).permute(1, 2, 3, 0) *\n 255).cpu().numpy().astype(np.uint8)\n try:\n writer = imageio.get_writer(\n video_path,\n fps=frame_rate,\n codec='libx264',\n quality=8,\n macro_block_size=1)\n for frame in video_frames:\n writer.append_data(frame)\n writer.close()\n print(video_path)\n except Exception as e:\n raise gr.Error(f\"Video save error: {e}\")\n if self.gallery_share:\n self.gallery_share_data.add(video_path)\n return self.gallery_share_data.get()\n else:\n return [video_path]\n def set_callbacks(self, **kwargs):\n self.gen_inputs = [\n self.output_gallery, self.src_video, self.src_mask,\n self.src_ref_image_1, self.src_ref_image_2, self.src_ref_image_3,\n self.prompt, self.negative_prompt, self.shift_scale,\n self.sample_steps, self.context_scale, self.guide_scale,\n self.infer_seed, self.output_height, self.output_width,\n self.frame_rate, self.num_frames\n ]\n self.gen_outputs = [self.output_gallery]\n self.generate_button.click(\n self.generate,\n inputs=self.gen_inputs,\n outputs=self.gen_outputs,\n queue=True)\n self.refresh_button.click(\n lambda x: self.gallery_share_data.get()\n if self.gallery_share else x,\n inputs=[self.output_gallery],\n outputs=[self.output_gallery])", "code_description": null, "fill_type": "CLASS_TYPE", "language_type": "python", "sub_task_type": null}, "context_code": [["/Wan2.1/gradio/fl2v_14B_singleGPU.py", "# Copyright 2024-2025 The Alibaba Wan Team Authors. All rights reserved.\nimport argparse\nimport gc\nimport os\nimport os.path as osp\nimport sys\nimport warnings\n\nimport gradio as gr\n\nwarnings.filterwarnings('ignore')\n\n# Model\nsys.path.insert(\n 0, os.path.sep.join(osp.realpath(__file__).split(os.path.sep)[:-2]))\nimport wan\nfrom wan.configs import MAX_AREA_CONFIGS, WAN_CONFIGS\nfrom wan.utils.prompt_extend import DashScopePromptExpander, QwenPromptExpander\nfrom wan.utils.utils import cache_video\n\n# Global Var\nprompt_expander = None\nwan_flf2v_720P = None\n\n\n# Button Func\ndef load_model(value):\n global wan_flf2v_720P\n\n if value == '------':\n print(\"No model loaded\")\n return '------'\n\n if value == '720P':\n if args.ckpt_dir_720p is None:\n print(\"Please specify the checkpoint directory for 720P model\")\n return '------'\n if wan_flf2v_720P is not None:\n pass\n else:\n gc.collect()\n\n print(\"load 14B-720P flf2v model...\", end='', flush=True)\n cfg = WAN_CONFIGS['flf2v-14B']\n wan_flf2v_720P = wan.WanFLF2V(\n config=cfg,\n checkpoint_dir=args.ckpt_dir_720p,\n device_id=0,\n rank=0,\n t5_fsdp=False,\n dit_fsdp=False,\n use_usp=False,\n )\n print(\"done\", flush=True)\n return '720P'\n return value\n\n\ndef prompt_enc(prompt, img_first, img_last, tar_lang):\n print('prompt extend...')\n if img_first is None or img_last is None:\n print('Please upload the first and last frames')\n return prompt\n global prompt_expander\n prompt_output = prompt_expander(\n prompt, image=[img_first, img_last], tar_lang=tar_lang.lower())\n if prompt_output.status == False:\n return prompt\n else:\n return prompt_output.prompt\n\n\ndef flf2v_generation(flf2vid_prompt, flf2vid_image_first, flf2vid_image_last,\n resolution, sd_steps, guide_scale, shift_scale, seed,\n n_prompt):\n\n if resolution == '------':\n print(\n 'Please specify the resolution ckpt dir or specify the resolution')\n return None\n\n else:\n if resolution == '720P':\n global wan_flf2v_720P\n video = wan_flf2v_720P.generate(\n flf2vid_prompt,\n flf2vid_image_first,\n flf2vid_image_last,\n max_area=MAX_AREA_CONFIGS['720*1280'],\n shift=shift_scale,\n sampling_steps=sd_steps,\n guide_scale=guide_scale,\n n_prompt=n_prompt,\n seed=seed,\n offload_model=True)\n pass\n else:\n print('Sorry, currently only 720P is supported.')\n return None\n\n cache_video(\n tensor=video[None],\n save_file=\"example.mp4\",\n fps=16,\n nrow=1,\n normalize=True,\n value_range=(-1, 1))\n\n return \"example.mp4\"\n\n\n# Interface\ndef gradio_interface():\n with gr.Blocks() as demo:\n gr.Markdown(\"\"\"\n