Sync local Space with Hub
Browse files
README.md
CHANGED
|
@@ -4,7 +4,7 @@ license: apache-2.0
|
|
| 4 |
|
| 5 |
# Pusa V1.0 Model
|
| 6 |
|
| 7 |
-
[Code Repository](https://github.com/Yaofang-Liu/Pusa-VidGen) | [Project Page](https://yaofang-liu.github.io/Pusa_Web/) |[Dataset](https://huggingface.co/datasets/RaphaelLiu/PusaV1_training) |[Model](https://huggingface.co/RaphaelLiu/PusaV1) | [Paper (Pusa V1.0)](https://github.com/Yaofang-Liu/Pusa-VidGen/blob/e99c3dcf866789a2db7fbe2686888ec398076a82/PusaV1/PusaV1.0_Report.pdf) | [Paper (FVDM)](https://arxiv.org/abs/2410.03160) | [Follow on X](https://x.com/stephenajason) | [Xiaohongshu](https://www.xiaohongshu.com/
|
| 8 |
|
| 9 |
## Overview
|
| 10 |
|
|
@@ -22,12 +22,6 @@ Pusa not only sets a new standard for image-to-video (I2V) generation but also *
|
|
| 22 |
|
| 23 |
Pusa V1.0, with only 10 inference steps, achieves state-of-the-art performance among open-source models. It surpasses its direct baseline, `Wan-I2V`, which was trained with vastly greater resources. Our model obtains a VBench-I2V total score of **87.32%**, outperforming `Wan-I2V`'s 86.86%.
|
| 24 |
|
| 25 |
-
|
| 26 |
-
## ✨ Key Features
|
| 27 |
-
|
| 28 |
-
- **Comprehensive Multi-task Support**:
|
| 29 |
-
// ... existing code ...
|
| 30 |
-
|
| 31 |
## ✨ Key Features
|
| 32 |
|
| 33 |
- **Comprehensive Multi-task Support**:
|
|
@@ -66,13 +60,17 @@ Pusa V1.0, with only 10 inference steps, achieves state-of-the-art performance a
|
|
| 66 |
### Download Weights and Recover The Checkpoint
|
| 67 |
|
| 68 |
**Option 1**: Use the Hugging Face CLI:
|
| 69 |
-
```
|
| 70 |
-
|
| 71 |
-
huggingface-cli
|
| 72 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 73 |
```
|
| 74 |
|
| 75 |
-
**Option 2**: Download directly from [Hugging Face](https://huggingface.co/RaphaelLiu/PusaV1) to your local machine.
|
| 76 |
|
| 77 |
## Related Work
|
| 78 |
|
|
|
|
| 4 |
|
| 5 |
# Pusa V1.0 Model
|
| 6 |
|
| 7 |
+
[Code Repository](https://github.com/Yaofang-Liu/Pusa-VidGen) | [Project Page](https://yaofang-liu.github.io/Pusa_Web/) |[Dataset](https://huggingface.co/datasets/RaphaelLiu/PusaV1_training) |[Model](https://huggingface.co/RaphaelLiu/PusaV1) | [Paper (Pusa V1.0)](https://github.com/Yaofang-Liu/Pusa-VidGen/blob/e99c3dcf866789a2db7fbe2686888ec398076a82/PusaV1/PusaV1.0_Report.pdf) | [Paper (FVDM)](https://arxiv.org/abs/2410.03160) | [Follow on X](https://x.com/stephenajason) | [Xiaohongshu](https://www.xiaohongshu.com/user/profile/5c6f928f0000000010015ca1?xsec_token=YBEf_x-s5bOBQIMJuNQvJ6H23Anwey1nnDgC9wiLyDHPU=&xsec_source=app_share&xhsshare=CopyLink&appuid=5c6f928f0000000010015ca1&apptime=1752622393&share_id=60f9a8041f974cb7ac5e3f0f161bf748)
|
| 8 |
|
| 9 |
## Overview
|
| 10 |
|
|
|
|
| 22 |
|
| 23 |
Pusa V1.0, with only 10 inference steps, achieves state-of-the-art performance among open-source models. It surpasses its direct baseline, `Wan-I2V`, which was trained with vastly greater resources. Our model obtains a VBench-I2V total score of **87.32%**, outperforming `Wan-I2V`'s 86.86%.
|
| 24 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 25 |
## ✨ Key Features
|
| 26 |
|
| 27 |
- **Comprehensive Multi-task Support**:
|
|
|
|
| 60 |
### Download Weights and Recover The Checkpoint
|
| 61 |
|
| 62 |
**Option 1**: Use the Hugging Face CLI:
|
| 63 |
+
```shell
|
| 64 |
+
# Make sure you are in the PusaV1 directory
|
| 65 |
+
# Install huggingface-cli if you don't have it
|
| 66 |
+
pip install -U "huggingface_hub[cli]"
|
| 67 |
+
huggingface-cli download RaphaelLiu/PusaV1 --local-dir ./model_zoo/PusaV1
|
| 68 |
+
|
| 69 |
+
# (Optional) Please download Wan2.1-T2V-14B to ./model_zoo/PusaV1 is you don't have it, if you have you can directly soft link it to ./model_zoo/PusaV1
|
| 70 |
+
huggingface-cli download Wan-AI/Wan2.1-T2V-14B --local-dir ./model_zoo/PusaV1
|
| 71 |
```
|
| 72 |
|
| 73 |
+
**Option 2**: Download directly `pusa_v1.pt` or `pusa_v1.safetensors` from [Hugging Face](https://huggingface.co/RaphaelLiu/PusaV1) to your local machine.
|
| 74 |
|
| 75 |
## Related Work
|
| 76 |
|