Video-to-Video
SeedVR
Iceclear commited on
Commit
bfbf2fe
·
verified ·
1 Parent(s): fe26bea

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -4
README.md CHANGED
@@ -2,7 +2,6 @@
2
  license: apache-2.0
3
  library_name: seedvr
4
  ---
5
-
6
  <div align="center">
7
  <img src="assets/seedvr_logo.png" alt="SeedVR" width="400"/>
8
  </div>
@@ -24,6 +23,18 @@ library_name: seedvr
24
  alt="SeedVR2 Paper on ArXiv"
25
  />
26
  </a>
 
 
 
 
 
 
 
 
 
 
 
 
27
  <a href="https://www.youtube.com/watch?v=tM8J-WhuAH0" target='_blank'>
28
  <img
29
  src="https://img.shields.io/badge/Demo%20Video-%23FF0000.svg?logo=YouTube&logoColor=white"
@@ -31,9 +42,8 @@ library_name: seedvr
31
  />
32
  </a>
33
  </p>
34
-
35
  >
36
- > Recent advances in diffusion-based video restoration (VR) demonstrate significant improvement in visual quality, yet yield a prohibitive computational cost during inference. While several distillation-based approaches have exhibited the potential of one-step image restoration, extending existing approaches to VR remains challenging and underexplored, due to the limited generation ability and poor temporal consistency, particularly when dealing with high-resolution video in real-world settings. In this work, we propose a one-step diffusion-based VR model, termed as AnonymousVR, which performs adversarial VR training against real data. To handle the challenging high-resolution VR within a single step, we introduce several enhancements to both model architecture and training procedures. Specifically, an adaptive window attention mechanism is proposed, where the window size is dynamically adjusted to fit the output resolutions, avoiding window inconsistency observed under high-resolution VR using window attention with a predefined window size. To stabilize and improve the adversarial post-training towards VR, we further verify the effectiveness of a series of losses, including a proposed feature matching loss without significantly sacrificing training efficiency. Extensive experiments show that AnonymousVR can achieve comparable or even better performance compared with existing VR approaches in a single step.
37
 
38
  <p align="center"><img src="assets/teaser.png" width="100%"></p>
39
 
@@ -51,7 +61,6 @@ library_name: seedvr
51
  booktitle={arXiv preprint arXiv:2506.05301},
52
  year={2025}
53
  }
54
-
55
  @inproceedings{wang2025seedvr,
56
  title={SeedVR: Seeding Infinity in Diffusion Transformer Towards Generic Video Restoration},
57
  author={Wang, Jianyi and Lin, Zhijie and Wei, Meng and Zhao, Yang and Yang, Ceyuan and Loy, Chen Change and Jiang, Lu},
 
2
  license: apache-2.0
3
  library_name: seedvr
4
  ---
 
5
  <div align="center">
6
  <img src="assets/seedvr_logo.png" alt="SeedVR" width="400"/>
7
  </div>
 
23
  alt="SeedVR2 Paper on ArXiv"
24
  />
25
  </a>
26
+ <a href="https://github.com/ByteDance-Seed/SeedVR">
27
+ <img
28
+ alt="Github" src="https://img.shields.io/badge/SeedVR2-Codebase-536af5?color=536af5&logo=github"
29
+ alt="SeedVR2 Codebase"
30
+ />
31
+ </a>
32
+ <a href="https://huggingface.co/models?other=seedvr">
33
+ <img
34
+ src="https://img.shields.io/badge/SeedVR2-Models-yellow?logo=huggingface&logoColor=yellow"
35
+ alt="SeedVR2 Models"
36
+ />
37
+ </a>
38
  <a href="https://www.youtube.com/watch?v=tM8J-WhuAH0" target='_blank'>
39
  <img
40
  src="https://img.shields.io/badge/Demo%20Video-%23FF0000.svg?logo=YouTube&logoColor=white"
 
42
  />
43
  </a>
44
  </p>
 
45
  >
46
+ > Recent advances in diffusion-based video restoration (VR) demonstrate significant improvement in visual quality, yet yield a prohibitive computational cost during inference. While several distillation-based approaches have exhibited the potential of one-step image restoration, extending existing approaches to VR remains challenging and underexplored, due to the limited generation ability and poor temporal consistency, particularly when dealing with high-resolution video in real-world settings. In this work, we propose a one-step diffusion-based VR model, termed as SeedVR2, which performs adversarial VR training against real data. To handle the challenging high-resolution VR within a single step, we introduce several enhancements to both model architecture and training procedures. Specifically, an adaptive window attention mechanism is proposed, where the window size is dynamically adjusted to fit the output resolutions, avoiding window inconsistency observed under high-resolution VR using window attention with a predefined window size. To stabilize and improve the adversarial post-training towards VR, we further verify the effectiveness of a series of losses, including a proposed feature matching loss without significantly sacrificing training efficiency. Extensive experiments show that SeedVR2 can achieve comparable or even better performance compared with existing VR approaches in a single step.
47
 
48
  <p align="center"><img src="assets/teaser.png" width="100%"></p>
49
 
 
61
  booktitle={arXiv preprint arXiv:2506.05301},
62
  year={2025}
63
  }
 
64
  @inproceedings{wang2025seedvr,
65
  title={SeedVR: Seeding Infinity in Diffusion Transformer Towards Generic Video Restoration},
66
  author={Wang, Jianyi and Lin, Zhijie and Wei, Meng and Zhao, Yang and Yang, Ceyuan and Loy, Chen Change and Jiang, Lu},