GrayShine commited on
Commit
4b44d77
·
verified ·
1 Parent(s): 0ee738c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -28,9 +28,9 @@
28
 
29
 
30
  ## 1. News
31
- - 2025-5-16:✨✨We release our 4 stages prograssive training code (supporting Huawei's NPU and NVIDIA's GPU). You can refer to [LVM/script/train](LVM/script/train) and [LVM/train](LVM/train) for detailed training information.
32
- - 2025-5-16:✨✨We release the inference code in [LVM/script/inference](LVM/script/inference) and [LVM/inference](LVM/inference).
33
- - 2025-5-16:🔥🔥We release the first version of Video-GPT. Model Weight: [Video-GPT](https://huggingface.co/GrayShine/Video-GPT)
34
 
35
 
36
  ## 2. Overview
@@ -41,7 +41,7 @@ Previous works on visual generation relies heavily on supervisory signals from t
41
 
42
  <!-- ![demo](https://github.com/zhuangshaobin/Video-GPT/tree/main/imgs/teaser.png) -->
43
  <p align="left">
44
- <img src="https://github.com/zhuangshaobin/Video-GPT/tree/main/imgs/teaser.png" alt="demo" width="640"/>
45
  </p>
46
 
47
  In addition, compared with the previous model architecture with many special designs for diffusion model (e.g., UNet, DiT, MM-DiT), we adopted the simplest vanilla transformer architecture. On the one hand, it is more conducive to the exploration of scaling law in the future. On the other hand, it is also more convenient for the community to follow up.
 
28
 
29
 
30
  ## 1. News
31
+ - 2025-5-21:✨✨We release our 4 stages prograssive training code (supporting Huawei's NPU and NVIDIA's GPU). You can refer to [LVM/script/train](LVM/script/train) and [LVM/train](LVM/train) for detailed training information.
32
+ - 2025-5-21:✨✨We release the inference code in [LVM/script/inference](LVM/script/inference) and [LVM/inference](LVM/inference).
33
+ - 2025-5-21:🔥🔥We release the first version of Video-GPT. Model Weight: [Video-GPT](https://huggingface.co/GrayShine/Video-GPT)
34
 
35
 
36
  ## 2. Overview
 
41
 
42
  <!-- ![demo](https://github.com/zhuangshaobin/Video-GPT/tree/main/imgs/teaser.png) -->
43
  <p align="left">
44
+ <img src="./imgs/teaser.png" alt="demo" width="640"/>
45
  </p>
46
 
47
  In addition, compared with the previous model architecture with many special designs for diffusion model (e.g., UNet, DiT, MM-DiT), we adopted the simplest vanilla transformer architecture. On the one hand, it is more conducive to the exploration of scaling law in the future. On the other hand, it is also more convenient for the community to follow up.