|
|
--- |
|
|
base_model: |
|
|
- alibaba-pai/Wan2.1-Fun-1.3B-Control |
|
|
- alibaba-pai/Wan2.1-Fun-14B-Control |
|
|
language: |
|
|
- en |
|
|
- zh |
|
|
license: apache-2.0 |
|
|
pipeline_tag: image-to-video |
|
|
--- |
|
|
|
|
|
# <center>π‘Lumen: Consistent Video Relighting and Harmonious Background Replacement with Video Generative Models </center> |
|
|
|
|
|
<div style="display: flex; justify-content: center; gap: 5px;"> |
|
|
<a href="https://lumen-relight.github.io"><img src="https://img.shields.io/badge/Project-Lumen-blue" alt="Project"></a> |
|
|
<a href="https://huggingface.co/papers/2508.12945"><img src="https://img.shields.io/badge/arXiv-Paper-red" alt="arXiv"></a> |
|
|
<a href="https://github.com/Kunbyte-AI/Lumen"><img src="https://img.shields.io/badge/GitHub-Code-black" alt="GitHub"></a> |
|
|
<a href="https://huggingface.co/Kunbyte/Lumen"><img src="https://img.shields.io/badge/π€%20HF-Model-yellow" alt="HuggingFace"></a> |
|
|
<a href="https://huggingface.co/spaces/Kunbyte/Lumen"><img src="https://img.shields.io/badge/π€%20HF-Space-yellow" alt="HuggingFace"></a> |
|
|
</div> |
|
|
|
|
|
π‘**Lumen** is an end-to-end video relighting framework developed on large-scale video generative models. It can relight the foreground and replace the background of a video based on flexible textual descriptions for instructing the control of lighting and background. |
|
|
|
|
|
## Introduction |
|
|
Video relighting aims to replace the background in videos while correspondingly adjusting the lighting in the foreground with harmonious blending. Lumen preserves the original properties of the foreground (e.g., albedo) and propagates consistent relighting across temporal frames. It is trained on a large-scale dataset featuring a mixture of realistic and synthetic videos, utilizing a domain-aware adapter to decouple the learning of relighting and domain appearance distribution. |
|
|
|
|
|
### Authors |
|
|
Jianshu Zeng, Yuxuan Liu, Yutong Feng, Chenxuan Miao, Zixiang Gao, Jiwang Qu, Jianzhang Zhang, Bin Wang, Kun Yuan. |
|
|
|
|
|
## π Quick Start |
|
|
|
|
|
This repository contains the weights of **Lumen**. For more instructions about how to use the model, please refer to the [official GitHub repository](https://github.com/Kunbyte-AI/Lumen). |
|
|
|
|
|
### Environment Setup |
|
|
```bash |
|
|
conda create -n lumen python=3.10 -y |
|
|
conda activate lumen |
|
|
pip install torch==2.4.0 torchvision==0.19.0 --index-url https://download.pytorch.org/whl/cu124 |
|
|
pip install -r requirements.txt |
|
|
``` |
|
|
|
|
|
### Inference |
|
|
After downloading the model weights and necessary base models (Wan2.1-Fun), you can run inference or the Gradio app: |
|
|
```bash |
|
|
# Run text-to-video inference |
|
|
python infer_t2v.py |
|
|
|
|
|
# Launch the Gradio demo |
|
|
python app_lumen.py |
|
|
``` |
|
|
|
|
|
## π Citation |
|
|
If you find our work helpful, please consider citing: |
|
|
```bibtex |
|
|
@article{zeng2025lumen, |
|
|
title={Lumen: Consistent Video Relighting and Harmonious Background Replacement with Video Generative Models}, |
|
|
author={Zeng, Jianshu and Liu, Yuxuan and Feng, Yutong and Miao, Chenxuan and Gao, Zixiang and Qu, Jiwang and Zhang, Jianzhang and Wang, Bin and Yuan, Kun}, |
|
|
journal={arXiv preprint arXiv:2508.12945}, |
|
|
year={2025}, |
|
|
url={https://arxiv.org/abs/2508.12945}, |
|
|
} |
|
|
``` |
|
|
|
|
|
## Acknowledgements |
|
|
We would like to thank the contributors to [DiffSynth-Studio](https://github.com/modelscope/DiffSynth-Studio), [VideoX-Fun](https://github.com/aigc-apps/VideoX-Fun), and the [Wan2.1](https://github.com/Wan-Video/Wan2.1) team for their open research and exploration. |