How to use joyfox/Wan2.2-I2V-KungFu with Diffusers:
pip install -U diffusers transformers accelerate
import torch from diffusers import DiffusionPipeline from diffusers.utils import load_image, export_to_video # switch to "mps" for apple devices pipe = DiffusionPipeline.from_pretrained("Wan-AI/Wan2.2-I2V-A14B", dtype=torch.bfloat16, device_map="cuda") pipe.load_lora_weights("joyfox/Wan2.2-I2V-KungFu") prompt = "一个小孩双脚直立,双臂灵活张开,时而抬手,然后转身朝向左边,时而踢腿,做着一系列打拳动作,wugong" input_image = load_image("https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/diffusers/guitar-man.png") image = pipe(image=input_image, prompt=prompt).frames[0] export_to_video(output, "output.mp4")
This LoRA is trained on the Wan2.2-I2V-A14B model.
The key trigger phrase is: wugong
wugong
For best results, use this prompt structure:
Base model