CineTrans: Learning to Generate Videos with Cinematic Transitions via Masked Diffusion Models

 β€‚  β€‚  β€‚  β€‚ 

Xiaoxue Wu1,2*, Bingjie Gao2,3, Yu Qiao2†, Yaohui Wang2†, Xinyuan Chen2†

1Fudan University 2Shanghai Artificial Intelligence Laboratory 3Shanghai Jiao Tong University

*Work done during internship at Shanghai AI Laboratory †Corresponding author

πŸ“₯ Installation

  1. Clone the Repository
git clone https://github.com/UknowSth/CineTrans.git
cd CineTrans
  1. Set up Environment
conda create -n cinetrans python==3.11.9
conda activate cinetrans

pip install torch==2.5.1 torchvision==0.20.1 --index-url https://download.pytorch.org/whl/cu118
pip install -r requirements.txt

πŸ€— Checkpoint

CineTrans-Unet

Download the required model weights and place them in the ckpt/ directory.

ckpt/
│── stable-diffusion-v1-4/
β”‚   β”œβ”€β”€ scheduler/
β”‚   β”œβ”€β”€ text_encoder/
β”‚   β”œβ”€β”€ tokenizer/  
β”‚   │── unet/
β”‚   └── vae_temporal_decoder/
│── checkpoint.pt
│── longclip-L.pt

For more inference details, please refer to the GitHub repository.


πŸ“‘ BiTeX

If you find CineTrans useful for your research and applications, please cite using this BibTeX:

@misc{wu2025cinetranslearninggeneratevideos,
      title={CineTrans: Learning to Generate Videos with Cinematic Transitions via Masked Diffusion Models}, 
      author={Xiaoxue Wu and Bingjie Gao and Yu Qiao and Yaohui Wang and Xinyuan Chen},
      year={2025},
      eprint={2508.11484},
      archivePrefix={arXiv},
      primaryClass={cs.CV},
      url={https://arxiv.org/abs/2508.11484}, 
}
Downloads last month
-
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Dataset used to train NumlockUknowSth/CineTrans-Unet

Paper for NumlockUknowSth/CineTrans-Unet