step t2v
#6
by
jabbamaster
- opened
- .gitattributes +0 -1
- README.md +3 -11
- assets/img_lightx2v.png +0 -3
.gitattributes
CHANGED
|
@@ -51,4 +51,3 @@ demos/output_lightx2v_wan_t2v_t03.mp4 filter=lfs diff=lfs merge=lfs -text
|
|
| 51 |
demos/output_lightx2v_wan_t2v_t06.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 52 |
demos/output_lightx2v_wan_t2v_t05.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 53 |
demos/output_lightx2v_wan_t2v_t01.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 54 |
-
assets/img_lightx2v.png filter=lfs diff=lfs merge=lfs -text
|
|
|
|
| 51 |
demos/output_lightx2v_wan_t2v_t06.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 52 |
demos/output_lightx2v_wan_t2v_t05.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 53 |
demos/output_lightx2v_wan_t2v_t01.mp4 filter=lfs diff=lfs merge=lfs -text
|
|
|
README.md
CHANGED
|
@@ -5,11 +5,7 @@ language:
|
|
| 5 |
- zh
|
| 6 |
pipeline_tag: text-to-video
|
| 7 |
tags:
|
| 8 |
-
|
| 9 |
-
- diffusion-single-file
|
| 10 |
-
- comfyui
|
| 11 |
-
- distillation
|
| 12 |
-
- LoRA
|
| 13 |
library_name: diffusers
|
| 14 |
inference:
|
| 15 |
parameters:
|
|
@@ -17,13 +13,9 @@ inference:
|
|
| 17 |
---
|
| 18 |
# Wan2.1-T2V-14B-StepDistill-CfgDistill
|
| 19 |
|
| 20 |
-
<p align="center">
|
| 21 |
-
<img src="assets/img_lightx2v.png" width=75%/>
|
| 22 |
-
<p>
|
| 23 |
-
|
| 24 |
## Overview
|
| 25 |
|
| 26 |
-
Wan2.1-T2V-14B-StepDistill-CfgDistill is an advanced text-to-video generation model built upon the Wan2.1-T2V-14B foundation. This approach allows the model to generate videos with significantly fewer inference steps (4 steps) and without classifier-free guidance, substantially reducing video generation time while maintaining high quality outputs.
|
| 27 |
|
| 28 |
## Video Demos
|
| 29 |
|
|
@@ -38,7 +30,7 @@ Our training code is modified based on the [Self-Forcing](https://github.com/gua
|
|
| 38 |
Our inference framework utilizes [lightx2v](https://github.com/ModelTC/lightx2v), a highly efficient inference engine that supports multiple models. This framework significantly accelerates the video generation process while maintaining high quality output.
|
| 39 |
|
| 40 |
```bash
|
| 41 |
-
bash scripts/
|
| 42 |
```
|
| 43 |
|
| 44 |
## License Agreement
|
|
|
|
| 5 |
- zh
|
| 6 |
pipeline_tag: text-to-video
|
| 7 |
tags:
|
| 8 |
+
- video generation
|
|
|
|
|
|
|
|
|
|
|
|
|
| 9 |
library_name: diffusers
|
| 10 |
inference:
|
| 11 |
parameters:
|
|
|
|
| 13 |
---
|
| 14 |
# Wan2.1-T2V-14B-StepDistill-CfgDistill
|
| 15 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 16 |
## Overview
|
| 17 |
|
| 18 |
+
Wan2.1-T2V-14B-StepDistill-CfgDistill is an advanced text-to-video generation model built upon the Wan2.1-T2V-14B foundation. This approach allows the model to generate videos with significantly fewer inference steps (4 or 8 steps) and without classifier-free guidance, substantially reducing video generation time while maintaining high quality outputs.
|
| 19 |
|
| 20 |
## Video Demos
|
| 21 |
|
|
|
|
| 30 |
Our inference framework utilizes [lightx2v](https://github.com/ModelTC/lightx2v), a highly efficient inference engine that supports multiple models. This framework significantly accelerates the video generation process while maintaining high quality output.
|
| 31 |
|
| 32 |
```bash
|
| 33 |
+
bash scripts/run_wan_t2v_distill.sh
|
| 34 |
```
|
| 35 |
|
| 36 |
## License Agreement
|
assets/img_lightx2v.png
DELETED
Git LFS Details
|