Update README.md
Browse files
README.md
CHANGED
|
@@ -70,7 +70,10 @@ or you can click <a href="https://github.com/PKU-YuanGroup/Helios-Page/blob/main
|
|
| 70 |
|
| 71 |
## 📣 Latest News!!
|
| 72 |
|
| 73 |
-
* `[2026.03.
|
|
|
|
|
|
|
|
|
|
| 74 |
* `[2026.03.04]` 🚀 Day-0 support for [Ascend-NPU](https://www.hiascend.com),with sincere gratitude to the Ascend Team for their support.
|
| 75 |
* `[2026.03.04]` 🚀 Day-0 support for [Diffusers](https://github.com/huggingface/diffusers/pull/13208),with special thanks to the HuggingFace Team for their support.
|
| 76 |
* `[2026.03.04]` 🚀 Day-0 support for [vLLM-Omni](https://github.com/vllm-project/vllm-omni/pull/1604),with heartfelt gratitude to the vLLM Team for their support.
|
|
@@ -83,9 +86,10 @@ or you can click <a href="https://github.com/PKU-YuanGroup/Helios-Page/blob/main
|
|
| 83 |
If your work has improved **Helios** and you would like more people to see it, please inform us.
|
| 84 |
|
| 85 |
* [Ascend-NPU](https://www.hiascend.com/): Developed by Huawei, this hardware is designed for efficient AI model training and inference, boosting performance in tasks like computer vision, natural language processing, and autonomous driving.
|
| 86 |
-
* [Diffusers](https://github.com/huggingface/diffusers): A popular library designed for working with diffusion models and other generative models in deep learning. It supports easy integration and manipulation of a wide range of generative models.
|
| 87 |
-
* [vLLM-Omni](https://github.com/vllm-project/vllm-omni): A fully disaggregated serving system for any-to-any models. vLLM-Omni breaks complex architectures into a stage-based graph, using a decoupled backend to maximize resource efficiency and throughput.
|
| 88 |
-
* [SGLang-Diffusion](https://github.com/sgl-project/sglang): An inference framework for accelerated image and video generation using diffusion models. It provides an end-to-end unified pipeline with optimized kernels and an efficient scheduler loop.
|
|
|
|
| 89 |
|
| 90 |
|
| 91 |
|
|
@@ -146,6 +150,28 @@ Before trying your own inputs, we highly recommend going through the sanity chec
|
|
| 146 |
| **V2V** | <video src="https://github.com/user-attachments/assets/420cb572-85c2-42d8-98d7-37b0bc24c844" controls width="240"></video> | <video src="https://github.com/user-attachments/assets/7d703fa6-dc1a-4138-a897-e58cfd9236d6" controls width="240"></video> | <video src="https://github.com/user-attachments/assets/45329c55-1a25-459c-bbf0-4e584ec5b23d" controls width="240"></video> |
|
| 147 |
|
| 148 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 149 |
### ✨ Diffusers Pipeline
|
| 150 |
|
| 151 |
Install diffusers from source:
|
|
|
|
| 70 |
|
| 71 |
## 📣 Latest News!!
|
| 72 |
|
| 73 |
+
* `[2026.03.06]` 🚀 [Cache-DiT](https://github.com/vipshop/cache-dit/pull/834) now supports Helios, it offers Fully Cache Acceleration and Parallelism support for Helios! Special thanks to the Cache-DiT Team for their amazing work.
|
| 74 |
+
* `[2026.03.06]` 🚀 We fix the Parallel Inference logits for Helios, and provide an example [here](#-parallel-inference-on-multiple-gpus). Thanks [Cache-DiT Team](https://github.com/vipshop/cache-dit/pull/836).
|
| 75 |
+
* `[2026.03.06]` 👋 We official release the [Gradio Demo](https://huggingface.co/spaces/BestWishYsh/Helios-14B-RealTime), welcome to try it.
|
| 76 |
+
* `[2026.03.05]` 👋 We are excited to announce the release of the Helios [technical report](https://arxiv.org/abs/2603.04379) on arXiv. We welcome discussions and feedback!
|
| 77 |
* `[2026.03.04]` 🚀 Day-0 support for [Ascend-NPU](https://www.hiascend.com),with sincere gratitude to the Ascend Team for their support.
|
| 78 |
* `[2026.03.04]` 🚀 Day-0 support for [Diffusers](https://github.com/huggingface/diffusers/pull/13208),with special thanks to the HuggingFace Team for their support.
|
| 79 |
* `[2026.03.04]` 🚀 Day-0 support for [vLLM-Omni](https://github.com/vllm-project/vllm-omni/pull/1604),with heartfelt gratitude to the vLLM Team for their support.
|
|
|
|
| 86 |
If your work has improved **Helios** and you would like more people to see it, please inform us.
|
| 87 |
|
| 88 |
* [Ascend-NPU](https://www.hiascend.com/): Developed by Huawei, this hardware is designed for efficient AI model training and inference, boosting performance in tasks like computer vision, natural language processing, and autonomous driving.
|
| 89 |
+
* [Diffusers](https://github.com/huggingface/diffusers/pull/13208): A popular library designed for working with diffusion models and other generative models in deep learning. It supports easy integration and manipulation of a wide range of generative models.
|
| 90 |
+
* [vLLM-Omni](https://github.com/vllm-project/vllm-omni/pull/1604): A fully disaggregated serving system for any-to-any models. vLLM-Omni breaks complex architectures into a stage-based graph, using a decoupled backend to maximize resource efficiency and throughput.
|
| 91 |
+
* [SGLang-Diffusion](https://github.com/sgl-project/sglang/pull/19782): An inference framework for accelerated image and video generation using diffusion models. It provides an end-to-end unified pipeline with optimized kernels and an efficient scheduler loop.
|
| 92 |
+
* [Cache-DiT](https://github.com/vipshop/cache-dit/pull/834): A PyTorch-native and Flexible Inference Engine with Hybrid Cache Acceleration and Parallelism for DiTs. It built on top of the Diffusers library and now supports nearly ALL DiTs from Diffusers.
|
| 93 |
|
| 94 |
|
| 95 |
|
|
|
|
| 150 |
| **V2V** | <video src="https://github.com/user-attachments/assets/420cb572-85c2-42d8-98d7-37b0bc24c844" controls width="240"></video> | <video src="https://github.com/user-attachments/assets/7d703fa6-dc1a-4138-a897-e58cfd9236d6" controls width="240"></video> | <video src="https://github.com/user-attachments/assets/45329c55-1a25-459c-bbf0-4e584ec5b23d" controls width="240"></video> |
|
| 151 |
|
| 152 |
|
| 153 |
+
### ✨ Parallel Inference on Multiple GPUs
|
| 154 |
+
For example, let's take Helios-Base with 2 GPUs.
|
| 155 |
+
|
| 156 |
+
<details>
|
| 157 |
+
<summary>Click to expand the code</summary>
|
| 158 |
+
|
| 159 |
+
```bash
|
| 160 |
+
CUDA_VISIBLE_DEVICES=0,1 torchrun --nproc_per_node 2 infer_helios.py \
|
| 161 |
+
--enable_parallelism \
|
| 162 |
+
--base_model_path "BestWishYsh/Helios-Base" \
|
| 163 |
+
--transformer_path "BestWishYsh/Helios-Base" \
|
| 164 |
+
--sample_type "t2v" \
|
| 165 |
+
--num_frames 99 \
|
| 166 |
+
--fps 24 \
|
| 167 |
+
--prompt "A vibrant tropical fish swimming gracefully among colorful coral reefs in a clear, turquoise ocean. The fish has bright blue and yellow scales with a small, distinctive orange spot on its side, its fins moving fluidly. The coral reefs are alive with a variety of marine life, including small schools of colorful fish and sea turtles gliding by. The water is crystal clear, allowing for a view of the sandy ocean floor below. The reef itself is adorned with a mix of hard and soft corals in shades of red, orange, and green. The photo captures the fish from a slightly elevated angle, emphasizing its lively movements and the vivid colors of its surroundings. A close-up shot with dynamic movement." \
|
| 168 |
+
--guidance_scale 5.0 \
|
| 169 |
+
--output_folder "./output_helios/helios-base"
|
| 170 |
+
```
|
| 171 |
+
|
| 172 |
+
</details>
|
| 173 |
+
|
| 174 |
+
|
| 175 |
### ✨ Diffusers Pipeline
|
| 176 |
|
| 177 |
Install diffusers from source:
|