Wan2.1 14B T2V LoRAs
Collection
A collection of Remade's Wan2.1 14B T2V LoRAs
• 20 items • Updated • 35
import torch
from diffusers import DiffusionPipeline
from diffusers.utils import export_to_video
# switch to "mps" for apple devices
pipe = DiffusionPipeline.from_pretrained("Wan-AI/Wan2.1-T2V-14B", dtype=torch.bfloat16, device_map="cuda")
pipe.load_lora_weights("Remade-AI/Zoom-Call")
prompt = "The video shows a [z00m_ca11] with four participants. In the top left box, a medieval knight in full armor adjusts his helmet. To his right, a pirate with a parrot on his shoulder drinks from a mug. In the bottom left, a scientist in a lab coat scribbles on a whiteboard. In the bottom right, an alien in a suit waves awkwardly."
output = pipe(prompt=prompt).frames[0]
export_to_video(output, "output.mp4")This LoRA is trained on the Wan2.1 14B T2V model and allows you to generate videos of Zoom calls featuring whatever character you want!
The key trigger phrase is: [z00m_ca11]
For prompting, check out the example prompts; this way of prompting seems to work very well.
This LoRA works with a modified version of Kijai's Wan Video Wrapper workflow. The main modification is adding a Wan LoRA node connected to the base model.
See the Downloads section above for the modified workflow.
The model weights are available in Safetensors format. See the Downloads section above.
Training was done using Diffusion Pipe for Training
Special thanks to Kijai for the ComfyUI Wan Video Wrapper and tdrussell for the training scripts!
Base model
Wan-AI/Wan2.1-T2V-14B