Improve model card: Add pipeline tag, library name, and consistent paper/Github links
Browse filesThis PR improves the model card for HuMo by:
- Adding `pipeline_tag: any-to-any` to help users discover the model for its broad multimodal video generation capabilities.
- Adding `library_name: diffusers` as evidence from `config.json` (`_diffusers_version`) indicates compatibility with the Diffusers library, enabling automated code snippets.
- Updating the arXiv badge and citation links to `https://arxiv.org/abs/2509.08519` for consistency with the paper's actual ID and the BibTeX.
- Adding a badge link to the GitHub repository (`https://github.com/Phantom-video/HuMo`) in the header for easy access to the code.
README.md
CHANGED
|
@@ -1,9 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
# HuMo: Human-Centric Video Generation via Collaborative Multi-Modal Conditioning
|
| 2 |
|
| 3 |
<div align="center">
|
| 4 |
|
| 5 |
-
[](https://phantom-video.github.io/HuMo/)
|
|
|
|
| 7 |
<a href="https://huggingface.co/bytedance-research/HuMo"><img src="https://img.shields.io/static/v1?label=%F0%9F%A4%97%20Hugging%20Face&message=Model&color=orange"></a>
|
| 8 |
</div>
|
| 9 |
|
|
@@ -111,15 +117,18 @@ Our work builds upon and is greatly inspired by several outstanding open-source
|
|
| 111 |
|
| 112 |
If HuMo is helpful, please help to ⭐ the repo.
|
| 113 |
|
| 114 |
-
If you find this project useful for your research, please consider citing our [paper](https://arxiv.org/abs/
|
| 115 |
|
| 116 |
### BibTeX
|
| 117 |
```bibtex
|
| 118 |
-
@
|
| 119 |
-
|
| 120 |
-
|
| 121 |
-
|
| 122 |
-
|
|
|
|
|
|
|
|
|
|
| 123 |
}
|
| 124 |
```
|
| 125 |
|
|
|
|
| 1 |
+
---
|
| 2 |
+
pipeline_tag: any-to-any
|
| 3 |
+
library_name: diffusers
|
| 4 |
+
---
|
| 5 |
+
|
| 6 |
# HuMo: Human-Centric Video Generation via Collaborative Multi-Modal Conditioning
|
| 7 |
|
| 8 |
<div align="center">
|
| 9 |
|
| 10 |
+
[](https://arxiv.org/abs/2509.08519)
|
| 11 |
[](https://phantom-video.github.io/HuMo/)
|
| 12 |
+
[](https://github.com/Phantom-video/HuMo)
|
| 13 |
<a href="https://huggingface.co/bytedance-research/HuMo"><img src="https://img.shields.io/static/v1?label=%F0%9F%A4%97%20Hugging%20Face&message=Model&color=orange"></a>
|
| 14 |
</div>
|
| 15 |
|
|
|
|
| 117 |
|
| 118 |
If HuMo is helpful, please help to ⭐ the repo.
|
| 119 |
|
| 120 |
+
If you find this project useful for your research, please consider citing our [paper](https://arxiv.org/abs/2509.08519).
|
| 121 |
|
| 122 |
### BibTeX
|
| 123 |
```bibtex
|
| 124 |
+
@misc{chen2025humo,
|
| 125 |
+
title={HuMo: Human-Centric Video Generation via Collaborative Multi-Modal Conditioning},
|
| 126 |
+
author={Liyang Chen and Tianxiang Ma and Jiawei Liu and Bingchuan Li and Zhuowei Chen and Lijie Liu and Xu He and Gen Li and Qian He and Zhiyong Wu},
|
| 127 |
+
year={2025},
|
| 128 |
+
eprint={2509.08519},
|
| 129 |
+
archivePrefix={arXiv},
|
| 130 |
+
primaryClass={cs.CV},
|
| 131 |
+
url={https://arxiv.org/abs/2509.08519},
|
| 132 |
}
|
| 133 |
```
|
| 134 |
|