--- tags: - text-to-image - lora - diffusers - template:diffusion-lora widget: - output: url: images/WanMoveLoraTesting_00003_.webp text: '-' base_model: 'bytedance-research/HuMo' instance_prompt: null license: apache-2.0 --- # HuMoveLora ## Model description A LoRA that tries to combine track controlled motion of [Wan-Move](https://github.com/ali-vilab/Wan-Move) with the human motion and speech sync features of [HuMo](https://huggingface.co/bytedance-research/HuMo). ## Usage - **Base model:** Wan HuMo-17B Model - **In ComfyUI:** Likely needs [WanExperiments](https://github.com/drozbay/WanExperiments) to allow for I2V capabilities. Chain the HuMo node with the WanMove node. **Recommended track strength: 1.5.** - It is *not recommended* to use HuMo reference or start image along with the WanMove start image. - [**Example ComfyUI WF**](./images/your_workflow_name.json) ## Version History | Version | Notes | |---------|-------| | v0.1 | Proof of concept release. Further tuning should improve simultaneous speech and motion control capabilities. | ## Download model [Download](/drozbay/HuMoveLora/tree/main) them in the Files & versions tab.