Improve model card: Add pipeline tag, library name, paper/code/project links, and citation
#4
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,9 +1,11 @@
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
| 3 |
tags:
|
| 4 |
-
|
| 5 |
-
|
| 6 |
-
|
|
|
|
|
|
|
| 7 |
---
|
| 8 |
|
| 9 |
<div align="center">
|
|
@@ -14,6 +16,8 @@ tags:
|
|
| 14 |
|
| 15 |
<h2>Expressive Portrait Image Animation for Live Streaming</h2>
|
| 16 |
|
|
|
|
|
|
|
| 17 |
[Zhiyuan Li<sup>1,2,3</sup>](https://huai-chang.github.io/) · [Chi-Man Pun<sup>1</sup>](https://cmpun.github.io/) 📪 · [Chen Fang<sup>2</sup>](http://fangchen.org/) · [Jue Wang<sup>2</sup>](https://scholar.google.com/citations?user=Bt4uDWMAAAAJ&hl=en) · [Xiaodong Cun<sup>3</sup>](https://vinthony.github.io/academic/) 📪
|
| 18 |
|
| 19 |
<sup>1</sup> University of Macau <sup>2</sup> [Dzine.ai](https://www.dzine.ai/) <sup>3</sup> [GVC Lab, Great Bay University](https://gvclab.github.io/)
|
|
@@ -125,11 +129,16 @@ python inference_online.py
|
|
| 125 |
```
|
| 126 |
then open `http://0.0.0.0:7860` in your browser. (*If `http://0.0.0.0:7860` does not work well, try `http://localhost:7860`)
|
| 127 |
|
| 128 |
-
<!-- ## 📋 Citation
|
| 129 |
-
If you find PersonaLive useful for your research, welcome to 🌟 this repo and cite our work using the following BibTeX:
|
| 130 |
-
```bibtex
|
| 131 |
-
|
| 132 |
-
``` -->
|
| 133 |
-
|
| 134 |
## ❤️ Acknowledgement
|
| 135 |
This code is mainly built upon [Moore-AnimateAnyone](https://github.com/MooreThreads/Moore-AnimateAnyone), [X-NeMo](https://byteaigc.github.io/X-Portrait2/), [StreamDiffusion](https://github.com/cumulo-autumn/StreamDiffusion), [RAIN](https://pscgylotti.github.io/pages/RAIN/) and [LivePortrait](https://github.com/KlingTeam/LivePortrait), thanks to their invaluable contributions.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
| 3 |
tags:
|
| 4 |
+
- portrait-animation
|
| 5 |
+
- real-time
|
| 6 |
+
- diffusion
|
| 7 |
+
pipeline_tag: image-to-video
|
| 8 |
+
library_name: diffusers
|
| 9 |
---
|
| 10 |
|
| 11 |
<div align="center">
|
|
|
|
| 16 |
|
| 17 |
<h2>Expressive Portrait Image Animation for Live Streaming</h2>
|
| 18 |
|
| 19 |
+
[**📚 Paper**](https://huggingface.co/papers/2512.11253) | [**💻 Code**](https://github.com/GVCLab/PersonaLive) | [**🏠 Project Page**](https://huai-chang.github.io/)
|
| 20 |
+
|
| 21 |
[Zhiyuan Li<sup>1,2,3</sup>](https://huai-chang.github.io/) · [Chi-Man Pun<sup>1</sup>](https://cmpun.github.io/) 📪 · [Chen Fang<sup>2</sup>](http://fangchen.org/) · [Jue Wang<sup>2</sup>](https://scholar.google.com/citations?user=Bt4uDWMAAAAJ&hl=en) · [Xiaodong Cun<sup>3</sup>](https://vinthony.github.io/academic/) 📪
|
| 22 |
|
| 23 |
<sup>1</sup> University of Macau <sup>2</sup> [Dzine.ai](https://www.dzine.ai/) <sup>3</sup> [GVC Lab, Great Bay University](https://gvclab.github.io/)
|
|
|
|
| 129 |
```
|
| 130 |
then open `http://0.0.0.0:7860` in your browser. (*If `http://0.0.0.0:7860` does not work well, try `http://localhost:7860`)
|
| 131 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 132 |
## ❤️ Acknowledgement
|
| 133 |
This code is mainly built upon [Moore-AnimateAnyone](https://github.com/MooreThreads/Moore-AnimateAnyone), [X-NeMo](https://byteaigc.github.io/X-Portrait2/), [StreamDiffusion](https://github.com/cumulo-autumn/StreamDiffusion), [RAIN](https://pscgylotti.github.io/pages/RAIN/) and [LivePortrait](https://github.com/KlingTeam/LivePortrait), thanks to their invaluable contributions.
|
| 134 |
+
|
| 135 |
+
## ⭐ Citation
|
| 136 |
+
If you find PersonaLive useful for your research, welcome to 🌟 this repo and cite our work using the following BibTeX:
|
| 137 |
+
```bibtex
|
| 138 |
+
@article{li2025personalive,
|
| 139 |
+
title={PersonaLive! Expressive Portrait Image Animation for Live Streaming},
|
| 140 |
+
author={Li, Zhiyuan and Pun, Chi-Man and Fang, Chen and Wang, Jue and Cun, Xiaodong},
|
| 141 |
+
journal={arXiv preprint arXiv:2512.11253},
|
| 142 |
+
year={2025}
|
| 143 |
+
}
|
| 144 |
+
```
|