OpenLipSync
This is a small repository containing all the required files to run inference for LatentSync1.5.
Installation
- Clone the repo
- On Debian-based systems, run
bash debian_setup.shfor both local and modal (remote inference) - For remote inference with Modal, you must first create the volume by running:
uv run modal run scripts/modal_download_extras.py
uv run modal run scripts/modal_download_models.py
Running Inference
Local Inference
Modify the inference.py file at the root of the directory (add the path of your video and file). Then run with:
uv run inference.py
Remote Inference
Modify the modal_lipsync_inference.py file at the root of the directory (add the path of your video and file). Then run with:
uv run modal run modal_lipsync_inference.py
Remote Inference with FastAPI Endpoints
Run:
uv run modal run modal_lipsync_serve.py
TODO:
- Add MuseTalk checkpoints
- Add LatentSync16 checkpoints