ryanontheinside commited on
Commit
746a605
·
verified ·
1 Parent(s): de2f148

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +60 -0
README.md ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ arxiv: "2602.14381"
3
+ tags:
4
+ - video-generation
5
+ - vace
6
+ - real-time
7
+ - autoregressive
8
+ - diffusion
9
+ - wan
10
+ license: apache-2.0
11
+ ---
12
+
13
+ # Adapting VACE for Real-Time Autoregressive Video Diffusion
14
+
15
+ This is the companion model card for the paper [Adapting VACE for Real-Time Autoregressive Video Diffusion](https://arxiv.org/abs/2602.14381).
16
+
17
+ ## Overview
18
+
19
+ This work presents modifications to [VACE](https://github.com/ali-vilab/VACE) that enable real-time autoregressive generation. The original VACE system uses bidirectional attention across full sequences, which is incompatible with streaming requirements. The key innovation moves reference frames from the diffusion latent space into a parallel conditioning pathway, maintaining fixed chunk sizes and KV caching needed for autoregressive models.
20
+
21
+ The adaptation leverages existing pretrained weights without retraining. Testing across 1.3B and 14B model scales shows structural control adds 20-30% latency overhead with minimal memory costs.
22
+
23
+ ## Real-Time Demo
24
+
25
+ Resolume Arena as live input into Scope via Spout:
26
+
27
+ <video src="https://huggingface.co/ryanontheinside/scope-vace/resolve/main/videos/resolume.mp4" controls autoplay loop muted></video>
28
+
29
+ ## VACE Control Examples
30
+
31
+ These comparisons show the adapted VACE conditioning across different control modes (corresponding to figures in the paper):
32
+
33
+ | Control Mode | Video |
34
+ |---|---|
35
+ | Depth | <video src="https://huggingface.co/ryanontheinside/scope-vace/resolve/main/videos/depth_comparison.mp4" controls loop muted width="400"></video> |
36
+ | Scribble | <video src="https://huggingface.co/ryanontheinside/scope-vace/resolve/main/videos/scribble_comparison.mp4" controls loop muted width="400"></video> |
37
+ | Optical Flow | <video src="https://huggingface.co/ryanontheinside/scope-vace/resolve/main/videos/optical_flow_comparison.mp4" controls loop muted width="400"></video> |
38
+ | Image-to-Video | <video src="https://huggingface.co/ryanontheinside/scope-vace/resolve/main/videos/i2v_comparison.mp4" controls loop muted width="400"></video> |
39
+ | Inpainting | <video src="https://huggingface.co/ryanontheinside/scope-vace/resolve/main/videos/inpainting_comparison.mp4" controls loop muted width="400"></video> |
40
+ | Outpainting | <video src="https://huggingface.co/ryanontheinside/scope-vace/resolve/main/videos/outpainting_comparison.mp4" controls loop muted width="400"></video> |
41
+ | Layout | <video src="https://huggingface.co/ryanontheinside/scope-vace/resolve/main/videos/layout_comparison.mp4" controls loop muted width="400"></video> |
42
+
43
+ ## Reference Implementation
44
+
45
+ The reference implementation is available in [Daydream Scope](https://github.com/daydreamlive/scope), a tool for running real-time, interactive generative AI video pipelines.
46
+
47
+ ## Author
48
+
49
+ [ryanontheinside.com](https://ryanontheinside.com)
50
+
51
+ ## Citation
52
+
53
+ ```bibtex
54
+ @article{fosdick2026adapting,
55
+ title={Adapting VACE for Real-Time Autoregressive Video Diffusion},
56
+ author={Fosdick, Ryan},
57
+ journal={arXiv preprint arXiv:2602.14381},
58
+ year={2026}
59
+ }
60
+ ```