Improve model card
#1
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,10 +1,11 @@
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
|
|
|
|
|
|
| 3 |
tags:
|
| 4 |
-
|
| 5 |
-
|
| 6 |
-
|
| 7 |
-
pipeline_tag: video-to-video
|
| 8 |
---
|
| 9 |
|
| 10 |
<div align="center">
|
|
@@ -16,7 +17,43 @@ pipeline_tag: video-to-video
|
|
| 16 |
|
| 17 |
**IWAIT 2026**
|
| 18 |
|
| 19 |
-
[](https://huggingface.co/papers/2601.11048)
|
| 21 |
+
[](https://github.com/tamaki-lab/M3DDM-Plus)
|
| 22 |
|
| 23 |
+
</div>
|
| 24 |
+
|
| 25 |
+
M3DDM+ provides a computationally efficient framework for video outpainting via latent diffusion modeling. It addresses significant quality degradation—such as spatial blur and temporal inconsistency—found in previous methods like M3DDM, especially under challenging scenarios with limited camera motion or large outpainting regions. By applying a uniform mask direction and width across all frames during training, M3DDM+ substantially improves visual fidelity and temporal coherence while maintaining computational efficiency.
|
| 26 |
+
|
| 27 |
+
## Sample Usage
|
| 28 |
+
|
| 29 |
+
For detailed environment setup and to download necessary pre-trained models, please refer to the [GitHub repository](https://github.com/tamaki-lab/M3DDM-Plus). Once set up, you can run inference using the following command:
|
| 30 |
+
|
| 31 |
+
```bash
|
| 32 |
+
CUDA_VISIBLE_DEVICES=0 python src/inference.py \
|
| 33 |
+
--input_video_path "path/to/input_video.mp4" \
|
| 34 |
+
--pretrained_sd_dir "stable-diffusion-v1-5" \
|
| 35 |
+
--video_outpainting_model_dir "MurakawaTakuya/M3DDM-Plus" \
|
| 36 |
+
--output_dir "path/to/output_directory" \
|
| 37 |
+
--target_ratio_list "1:1" \
|
| 38 |
+
--output_size 256
|
| 39 |
+
```
|
| 40 |
+
|
| 41 |
+
## Citation
|
| 42 |
+
|
| 43 |
+
If our work is helpful, please help to ⭐ the repo.
|
| 44 |
+
|
| 45 |
+
Please consider citing our paper if you found our work interesting and useful.
|
| 46 |
+
|
| 47 |
+
```bibtex
|
| 48 |
+
@article{murakawa2026m3ddmplus,
|
| 49 |
+
title={M3DDM+: An improved video outpainting by a modified masking strategy},
|
| 50 |
+
author={Murakawa, Takuya and Fukuzawa, Takumi and Ding, Ning and Tamaki, Toru},
|
| 51 |
+
journal={International Workshop on Advanced Image Technology (IWAIT)},
|
| 52 |
+
year={2026}
|
| 53 |
+
}
|
| 54 |
+
```
|
| 55 |
+
|
| 56 |
+
## Acknowledgement
|
| 57 |
+
|
| 58 |
+
The inference and pipeline code is based on published code of [M3DDM-Video-Outpainting](https://github.com/alimama-creative/M3DDM-Video-Outpainting).
|
| 59 |
+
The training and evaluation code was reproduced based on the M3DDM paper as it isn't published, and modified for our proposed method.
|