Update README.md
Browse files
README.md
CHANGED
|
@@ -28,9 +28,9 @@
|
|
| 28 |
|
| 29 |
|
| 30 |
## 1. News
|
| 31 |
-
- 2025-5-
|
| 32 |
-
- 2025-5-
|
| 33 |
-
- 2025-5-
|
| 34 |
|
| 35 |
|
| 36 |
## 2. Overview
|
|
@@ -41,7 +41,7 @@ Previous works on visual generation relies heavily on supervisory signals from t
|
|
| 41 |
|
| 42 |
<!--  -->
|
| 43 |
<p align="left">
|
| 44 |
-
<img src="
|
| 45 |
</p>
|
| 46 |
|
| 47 |
In addition, compared with the previous model architecture with many special designs for diffusion model (e.g., UNet, DiT, MM-DiT), we adopted the simplest vanilla transformer architecture. On the one hand, it is more conducive to the exploration of scaling law in the future. On the other hand, it is also more convenient for the community to follow up.
|
|
|
|
| 28 |
|
| 29 |
|
| 30 |
## 1. News
|
| 31 |
+
- 2025-5-21:✨✨We release our 4 stages prograssive training code (supporting Huawei's NPU and NVIDIA's GPU). You can refer to [LVM/script/train](LVM/script/train) and [LVM/train](LVM/train) for detailed training information.
|
| 32 |
+
- 2025-5-21:✨✨We release the inference code in [LVM/script/inference](LVM/script/inference) and [LVM/inference](LVM/inference).
|
| 33 |
+
- 2025-5-21:🔥🔥We release the first version of Video-GPT. Model Weight: [Video-GPT](https://huggingface.co/GrayShine/Video-GPT)
|
| 34 |
|
| 35 |
|
| 36 |
## 2. Overview
|
|
|
|
| 41 |
|
| 42 |
<!--  -->
|
| 43 |
<p align="left">
|
| 44 |
+
<img src="./imgs/teaser.png" alt="demo" width="640"/>
|
| 45 |
</p>
|
| 46 |
|
| 47 |
In addition, compared with the previous model architecture with many special designs for diffusion model (e.g., UNet, DiT, MM-DiT), we adopted the simplest vanilla transformer architecture. On the one hand, it is more conducive to the exploration of scaling law in the future. On the other hand, it is also more convenient for the community to follow up.
|