Improve model card

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +44 -7
README.md CHANGED
@@ -1,10 +1,11 @@
1
  ---
2
  license: apache-2.0
 
 
3
  tags:
4
- - video-outpainting
5
- - diffusion
6
- - video-generation
7
- pipeline_tag: video-to-video
8
  ---
9
 
10
  <div align="center">
@@ -16,7 +17,43 @@ pipeline_tag: video-to-video
16
 
17
  **IWAIT 2026**
18
 
19
- [![arXiv](https://img.shields.io/badge/arXiv-aaa-b31b1b.svg)]()
20
- [![Project Page](https://img.shields.io/badge/Project_Page-Website-blue)]()
21
 
22
- TBW
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ pipeline_tag: image-to-video
4
+ library_name: diffusers
5
  tags:
6
+ - video-outpainting
7
+ - diffusion
8
+ - video-generation
 
9
  ---
10
 
11
  <div align="center">
 
17
 
18
  **IWAIT 2026**
19
 
20
+ [![arXiv](https://img.shields.io/badge/arXiv-2601.11048-b31b1b.svg)](https://huggingface.co/papers/2601.11048)
21
+ [![GitHub](https://img.shields.io/badge/GitHub-Repo-blue)](https://github.com/tamaki-lab/M3DDM-Plus)
22
 
23
+ </div>
24
+
25
+ M3DDM+ provides a computationally efficient framework for video outpainting via latent diffusion modeling. It addresses significant quality degradation—such as spatial blur and temporal inconsistency—found in previous methods like M3DDM, especially under challenging scenarios with limited camera motion or large outpainting regions. By applying a uniform mask direction and width across all frames during training, M3DDM+ substantially improves visual fidelity and temporal coherence while maintaining computational efficiency.
26
+
27
+ ## Sample Usage
28
+
29
+ For detailed environment setup and to download necessary pre-trained models, please refer to the [GitHub repository](https://github.com/tamaki-lab/M3DDM-Plus). Once set up, you can run inference using the following command:
30
+
31
+ ```bash
32
+ CUDA_VISIBLE_DEVICES=0 python src/inference.py \
33
+ --input_video_path "path/to/input_video.mp4" \
34
+ --pretrained_sd_dir "stable-diffusion-v1-5" \
35
+ --video_outpainting_model_dir "MurakawaTakuya/M3DDM-Plus" \
36
+ --output_dir "path/to/output_directory" \
37
+ --target_ratio_list "1:1" \
38
+ --output_size 256
39
+ ```
40
+
41
+ ## Citation
42
+
43
+ If our work is helpful, please help to ⭐ the repo.
44
+
45
+ Please consider citing our paper if you found our work interesting and useful.
46
+
47
+ ```bibtex
48
+ @article{murakawa2026m3ddmplus,
49
+ title={M3DDM+: An improved video outpainting by a modified masking strategy},
50
+ author={Murakawa, Takuya and Fukuzawa, Takumi and Ding, Ning and Tamaki, Toru},
51
+ journal={International Workshop on Advanced Image Technology (IWAIT)},
52
+ year={2026}
53
+ }
54
+ ```
55
+
56
+ ## Acknowledgement
57
+
58
+ The inference and pipeline code is based on published code of [M3DDM-Video-Outpainting](https://github.com/alimama-creative/M3DDM-Video-Outpainting).
59
+ The training and evaluation code was reproduced based on the M3DDM paper as it isn't published, and modified for our proposed method.