diff --git "a/2025/VAFlow_ Video-to-Audio Generation with Cross-Modality Flow Matching/layout.json" "b/2025/VAFlow_ Video-to-Audio Generation with Cross-Modality Flow Matching/layout.json" new file mode 100644--- /dev/null +++ "b/2025/VAFlow_ Video-to-Audio Generation with Cross-Modality Flow Matching/layout.json" @@ -0,0 +1,8869 @@ +{ + "pdf_info": [ + { + "para_blocks": [ + { + "bbox": [ + 78, + 103, + 533, + 121 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 78, + 103, + 533, + 121 + ], + "spans": [ + { + "bbox": [ + 78, + 103, + 533, + 121 + ], + "type": "text", + "content": "VAFlow: Video-to-Audio Generation with Cross-Modality Flow Matching" + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 69, + 141, + 543, + 185 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 69, + 141, + 543, + 185 + ], + "spans": [ + { + "bbox": [ + 69, + 141, + 543, + 185 + ], + "type": "text", + "content": "Xihua Wang" + }, + { + "bbox": [ + 69, + 141, + 543, + 185 + ], + "type": "inline_equation", + "content": "^{1}" + }, + { + "bbox": [ + 69, + 141, + 543, + 185 + ], + "type": "text", + "content": " Xin Cheng" + }, + { + "bbox": [ + 69, + 141, + 543, + 185 + ], + "type": "inline_equation", + "content": "^{1}" + }, + { + "bbox": [ + 69, + 141, + 543, + 185 + ], + "type": "text", + "content": " Yuyue Wang" + }, + { + "bbox": [ + 69, + 141, + 543, + 185 + ], + "type": "inline_equation", + "content": "^{1}" + }, + { + "bbox": [ + 69, + 141, + 543, + 185 + ], + "type": "text", + "content": " Ruihua Song" + }, + { + "bbox": [ + 69, + 141, + 543, + 185 + ], + "type": "inline_equation", + "content": "^{1*}" + }, + { + "bbox": [ + 69, + 141, + 543, + 185 + ], + "type": "text", + "content": " Yunfeng Wang" + }, + { + "bbox": [ + 69, + 141, + 543, + 185 + ], + "type": "inline_equation", + "content": "^{2}" + }, + { + "bbox": [ + 69, + 141, + 543, + 185 + ], + "type": "inline_equation", + "content": "^{1}" + }, + { + "bbox": [ + 69, + 141, + 543, + 185 + ], + "type": "text", + "content": "Gaoling School of Artificial Intelligence, Renmin University of China \nxihuaw@ruc.edu.cn, songruihua.bloon@outlook.com" + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 152, + 213, + 200, + 225 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 152, + 213, + 200, + 225 + ], + "spans": [ + { + "bbox": [ + 152, + 213, + 200, + 225 + ], + "type": "text", + "content": "Abstract" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 53, + 239, + 297, + 515 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 53, + 239, + 297, + 515 + ], + "spans": [ + { + "bbox": [ + 53, + 239, + 297, + 515 + ], + "type": "text", + "content": "Video-to-audio (V2A) generation aims to synthesize temporally aligned, realistic sounds for silent videos, a critical capability for immersive multimedia applications. Current V2A methods, predominantly based on diffusion or flow models, rely on suboptimal noise-to-audio paradigms that entangle cross-modal mappings with stochastic priors, resulting in inefficient training and convoluted transport paths. We propose VAFlow, a novel flow-based framework that directly models the video-to-audio transformation, eliminating reliance on noise priors. To address modality discrepancies, we employ an alignment variational autoencoder that compresses heterogeneous video features into audio-aligned latent spaces while preserving spatiotemporal semantics. By retaining cross-attention mechanisms between video features and flow blocks, our architecture enables classifier-free guidance within video source-driven generation. Without external data or complex training tricks, VAFlow achieves state-of-the-art performance on VGGSound benchmark, surpassing even textaugmented models in audio fidelity, diversity, and distribution alignment. This work establishes a new paradigm for V2A generation with a direct and effective video-to-audio transformation via flow matching." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 56, + 540, + 135, + 553 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 540, + 135, + 553 + ], + "spans": [ + { + "bbox": [ + 56, + 540, + 135, + 553 + ], + "type": "text", + "content": "1. Introduction" + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 55, + 560, + 296, + 693 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 560, + 296, + 693 + ], + "spans": [ + { + "bbox": [ + 55, + 560, + 296, + 693 + ], + "type": "text", + "content": "Video-to-audio (V2A) generation plays a vital role in multimedia content generation, including Foley for films and AI-generated silent videos. In recent years, this task has attracted increasing attention within the generative community. Early V2A approaches [15, 24, 26, 29, 34] adopted language modeling strategies by discretizing audio into tokens, and used parallel mask-prediction models (BERT-like) or token-by-token autoregressive models (GPT-like) borrowed from NLP community as shown in Figure 1. However, these methods, reliant on discrete tokens, incur inherent information losses by the discretization process." + } + ] + } + ], + "index": 7 + }, + { + "type": "image", + "bbox": [ + 316, + 211, + 555, + 375 + ], + "blocks": [ + { + "bbox": [ + 316, + 211, + 555, + 375 + ], + "lines": [ + { + "bbox": [ + 316, + 211, + 555, + 375 + ], + "spans": [ + { + "bbox": [ + 316, + 211, + 555, + 375 + ], + "type": "image", + "image_path": "00b726f5ccb3264f0217ff01f4838780de47120879ff2588490bb01d049831e2.jpg" + } + ] + } + ], + "index": 8, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 313, + 384, + 556, + 484 + ], + "lines": [ + { + "bbox": [ + 313, + 384, + 556, + 484 + ], + "spans": [ + { + "bbox": [ + 313, + 384, + 556, + 484 + ], + "type": "text", + "content": "Figure 1. Four different V2A paradigms, along with state-of-the-art performance under each. The first row shows discrete token-based approaches (mask-prediction and autoregressive based generation), which are limited by information loss during tokenization. The second row presents methods generating continuous data. Previous diffusion and flow methods (bottom-left) model the noise-to-audio mapping conditioning on video. We find this approach suboptimal and demonstrate that directly modeling the video-to-audio transformation (bottom-right) yields better results." + } + ] + } + ], + "index": 9, + "angle": 0, + "type": "image_caption" + } + ], + "index": 8 + }, + { + "bbox": [ + 313, + 508, + 556, + 652 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 508, + 556, + 652 + ], + "spans": [ + { + "bbox": [ + 313, + 508, + 556, + 652 + ], + "type": "text", + "content": "Recent advancements in Diffusion and Flow Matching (FM) models in image [4, 10, 33] and video [3, 31] generation have spurred similar approaches for V2A [9, 25, 41, 43, 45]. Generally, these methods model V2A as a conditional generation task, transforming Gaussian noise conditioned on video into audio, as shown in Figure 1 (bottom-left). The generation process typically involves two steps: (1) sampling a random noise from Gaussian prior and (2) using this noise and video condition as input to iteratively denoise the noise into audio. This approach avoids discrete audio tokens, directly fitting the continuous distribution of audio data, and currently holds state-of-the-art results." + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 313, + 653, + 556, + 715 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 653, + 556, + 715 + ], + "spans": [ + { + "bbox": [ + 313, + 653, + 556, + 715 + ], + "type": "text", + "content": "However, existing diffusion/FM-based V2A methods focus solely on the second step, i.e., denoising Gaussian noise into audio, assuming the first step is trivial. Even state-of-the-art methods [40, 43] begin with Gaussian noise, but we find that this approach is not optimal. Unlike text-to" + } + ] + } + ], + "index": 11 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 65, + 2, + 111, + 34 + ], + "type": "header", + "angle": 0, + "lines": [ + { + "bbox": [ + 65, + 2, + 111, + 34 + ], + "spans": [ + { + "bbox": [ + 65, + 2, + 111, + 34 + ], + "type": "text", + "content": "CVF" + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 146, + 0, + 494, + 37 + ], + "type": "header", + "angle": 0, + "lines": [ + { + "bbox": [ + 146, + 0, + 494, + 37 + ], + "spans": [ + { + "bbox": [ + 146, + 0, + 494, + 37 + ], + "type": "text", + "content": "This ICCV paper is the Open Access version, provided by the Computer Vision Foundation. Except for this watermark, it is identical to the accepted version; the final published version of the proceedings is available on IEEE Xplore." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 67, + 703, + 145, + 713 + ], + "type": "page_footnote", + "angle": 0, + "lines": [ + { + "bbox": [ + 67, + 703, + 145, + 713 + ], + "spans": [ + { + "bbox": [ + 67, + 703, + 145, + 713 + ], + "type": "text", + "content": "*Corresponding author." + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "spans": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "type": "text", + "content": "11777" + } + ] + } + ], + "index": 13 + } + ], + "page_size": [ + 612, + 792 + ], + "page_idx": 0 + }, + { + "para_blocks": [ + { + "bbox": [ + 55, + 72, + 294, + 228 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 72, + 294, + 228 + ], + "spans": [ + { + "bbox": [ + 55, + 72, + 294, + 228 + ], + "type": "text", + "content": "image or text-to-video tasks, where textual conditions are abstract and highly defined by human, the video conditions in V2A are more complex and require both spatial and temporal understanding. The generation goal also demands a high degree of alignment between the generated audio and the video conditions, not just semantically, but also temporally. There are two inherent difficulties: (1) different conditions on the same noise require the model to generate different audio outputs, demanding advanced condition processing [8, 41, 46], and (2) noise variations under the same condition lead to unstable generation results, prompting research into methods for noise control [44], such as finding a \"golden noise\" for consistent results [47]." + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 55, + 230, + 294, + 337 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 230, + 294, + 337 + ], + "spans": [ + { + "bbox": [ + 55, + 230, + 294, + 337 + ], + "type": "text", + "content": "In contrast, we propose a new perspective by shifting focus to the first step of the generation process. We question: If there are better priors than Gaussian noise? Given that many diffusion and FM models dedicate significant effort to integrating video conditions into the denoising process for better generation performance, we further question: If V2A can work better when directly transforming the video distribution to audio rather than from noise? Fortunately, both answers are affirmative." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 55, + 340, + 295, + 591 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 340, + 295, + 591 + ], + "spans": [ + { + "bbox": [ + 55, + 340, + 295, + 591 + ], + "type": "text", + "content": "We introduce VAFlow, a novel framework for V2A generation, which directly denoises from the video space to the audio space. To implement this video-to-audio paradigm, we adopt the FM framework, as it theoretically supports any source distribution. There are three technical challenges for this implementation: (1) aligning source and target distributions: video and audio feature spaces typically have different time and spatial resolutions, yet FM requires the source and target distributions to be aligned (the same shape). (2) uncertainty of the feature type: identifying the optimal video representations/feature types as the source distribution. (3) conditioning mechanism: exploring additional video information integration ways beyond encoding it in the source distribution. In VAFlow, we first employ an alignment VAE to adjust video features to match the audio latent space. It then explores three different video feature types (semantic CLIP [32], spatiotemporal CAVP [25], reconstruction-focused VidTok [38]) to systematically investigate source video information. Experiments show that using video features both as a source distribution and as conditioning information yields the best results for VAFlow." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 55, + 594, + 295, + 714 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 594, + 295, + 714 + ], + "spans": [ + { + "bbox": [ + 55, + 594, + 295, + 714 + ], + "type": "text", + "content": "Moreover, the recent success of large language models (LLMs) [13, 28] has highlighted the effectiveness of scaling laws, and various generative works consequently explore similar scaling properties in each domain [30, 42]. However, existing V2A models have not demonstrated clear scaling trends, where increasing model size does not consistently achieve improvement [8, 43]. In contrast, our experiments show that VAFlow exhibits clear scaling property, with model size increases leading to consistent improvements in generation quality, further confirming its potential." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 313, + 72, + 553, + 156 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 72, + 553, + 156 + ], + "spans": [ + { + "bbox": [ + 313, + 72, + 553, + 156 + ], + "type": "text", + "content": "Without elaborate training tricks or external data (e.g., text-audio pairs), VAFlow achieves state-of-the-art results on standard benchmarks (VGGSound), outperforming even text-augmented models in generation fidelity, diversity, and cross-modality alignment with input video. Playable audi-los generated with VAFlow can be accessed at the demo page: https://vaflow.github.io/demo." + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 313, + 168, + 400, + 181 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 168, + 400, + 181 + ], + "spans": [ + { + "bbox": [ + 313, + 168, + 400, + 181 + ], + "type": "text", + "content": "2. Related Work" + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 313, + 189, + 511, + 201 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 189, + 511, + 201 + ], + "spans": [ + { + "bbox": [ + 313, + 189, + 511, + 201 + ], + "type": "text", + "content": "2.1. Flow Matching and Diffusion Models" + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 313, + 206, + 553, + 434 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 206, + 553, + 434 + ], + "spans": [ + { + "bbox": [ + 313, + 206, + 553, + 434 + ], + "type": "text", + "content": "Diffusion models [17, 27, 35] reconstruct data distributions by iteratively denoising Gaussian-sampled latents, governmented by various designed schedulers achieving state-of-the-art performance in multimodal generation tasks. Unlike diffusions, flow-based models [1, 20, 22, 23] establish deterministic mappings between source (e.g., Gaussian) and target distributions through invertible neural transformations, with recent flow matching optimizing vector fields through optimal transport theory, enabling direct learning of straight transport paths between source and target distributions, thereby reducing training instability and inference steps. While both paradigms typically rely on Gaussian priors as their source distribution, existing video-to-audio systems on both paradigms model noise-to-audio generation conditioned on video features. In contrast, our approach directly aligns video latent spaces with target audio distributions via linear transformations under a flow matching framework, bypassing iterative noise modeling, preserving efficiency, and ensuring precise frame-to-frame alignment." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 313, + 443, + 464, + 455 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 443, + 464, + 455 + ], + "spans": [ + { + "bbox": [ + 313, + 443, + 464, + 455 + ], + "type": "text", + "content": "2.2. Video-to-Audio Generation" + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 313, + 461, + 553, + 700 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 461, + 553, + 700 + ], + "spans": [ + { + "bbox": [ + 313, + 461, + 553, + 700 + ], + "type": "text", + "content": "Framework. Recent advances in generative models have driven significant progress in video-to-audio (V2A) synthesis. Current V2A methods fall into three main paradigms: (1) Autoregressive (AR) models [15, 26, 34] discretize audio into token (codec) sequences and utilize video features as prefixes to model unified video-audio sequences. While straightforward, their token-by-token generation process results in quadratic inference time scaling, limiting efficiency for long audio sequences. (2) Mask-based generation paradigms [24, 29, 37] also employ discrete tokens but decode audio in parallel by predicting masked tokens. Video features serve as the condition for mask-to-audio prediction. However, both AR and mask-based methods suffer from an inherent upper bound of fidelity due to lossy compression from audio discretization. (3) Flow/diffusion-based methods [7-9, 25, 40, 41, 43, 45, 46] leverage continuous representations and currently achieve state-of-the-art V2A performance [8, 40, 43]. These techniques iteratively transform noise into target audio distributions while conditioning on video features to steer the denoising trajectory." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 313, + 701, + 553, + 713 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 701, + 553, + 713 + ], + "spans": [ + { + "bbox": [ + 313, + 701, + 553, + 713 + ], + "type": "text", + "content": "Video Conditional Modeling. Unlike typical conditioned" + } + ] + } + ], + "index": 10 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "spans": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "type": "text", + "content": "11778" + } + ] + } + ], + "index": 11 + } + ], + "page_size": [ + 612, + 792 + ], + "page_idx": 1 + }, + { + "para_blocks": [ + { + "bbox": [ + 55, + 72, + 296, + 264 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 72, + 296, + 264 + ], + "spans": [ + { + "bbox": [ + 55, + 72, + 296, + 264 + ], + "type": "text", + "content": "generation tasks like text-to-image [33], V2A involves fully aligned sequential correspondences between video and audio. Existing methods address this through various strategies: some AR-based methods investigate various attention mechanisms and masking schemes for temporally sequential alignment [26]; some mask-based approaches integrate hybrid ways to fuse video features into the mask-prediction process to synchronize audio generation [29]; and some flow/diffusion-based methods augment their pipelines with auxiliary modules (e.g., loudness or audio layout predictors [41, 46]) to enforce temporal consistency. In contrast, our approach builds on flow-based techniques by discarding the conventional noise-to-audio paradigm and complex video-conditioning designs. Instead, we propose a more intuitive method that directly transforms the video temporal space to the audio temporal space via a flow transport." + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 55, + 274, + 115, + 287 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 274, + 115, + 287 + ], + "spans": [ + { + "bbox": [ + 55, + 274, + 115, + 287 + ], + "type": "text", + "content": "3. Method:" + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 55, + 293, + 214, + 306 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 293, + 214, + 306 + ], + "spans": [ + { + "bbox": [ + 55, + 293, + 214, + 306 + ], + "type": "text", + "content": "3.1. Preliminary: Flow Matching" + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 55, + 311, + 296, + 373 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 311, + 296, + 373 + ], + "spans": [ + { + "bbox": [ + 55, + 311, + 296, + 373 + ], + "type": "text", + "content": "Let " + }, + { + "bbox": [ + 55, + 311, + 296, + 373 + ], + "type": "inline_equation", + "content": "x_0 \\sim p(x_0)" + }, + { + "bbox": [ + 55, + 311, + 296, + 373 + ], + "type": "text", + "content": " and " + }, + { + "bbox": [ + 55, + 311, + 296, + 373 + ], + "type": "inline_equation", + "content": "x_1 \\sim p(x_1)" + }, + { + "bbox": [ + 55, + 311, + 296, + 373 + ], + "type": "text", + "content": " denote samples from source and target distributions, respectively, with " + }, + { + "bbox": [ + 55, + 311, + 296, + 373 + ], + "type": "inline_equation", + "content": "c" + }, + { + "bbox": [ + 55, + 311, + 296, + 373 + ], + "type": "text", + "content": " representing an optional conditioning signal. Flow Matching (FM) [20, 23] establishes an inter-distribution transport process governed by the ordinary differential equation (ODE):" + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 128, + 380, + 295, + 394 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 128, + 380, + 295, + 394 + ], + "spans": [ + { + "bbox": [ + 128, + 380, + 295, + 394 + ], + "type": "interline_equation", + "content": "d x (t) = v _ {\\theta} \\left(x _ {t}, t, c\\right) d t, \\tag {1}", + "image_path": "e6878073826b5f115ea9385be6517d07d4032dbb2f5242eb16bdb66e9e73b342.jpg" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 55, + 401, + 296, + 475 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 401, + 296, + 475 + ], + "spans": [ + { + "bbox": [ + 55, + 401, + 296, + 475 + ], + "type": "text", + "content": "where " + }, + { + "bbox": [ + 55, + 401, + 296, + 475 + ], + "type": "inline_equation", + "content": "t \\in [0,1]" + }, + { + "bbox": [ + 55, + 401, + 296, + 475 + ], + "type": "text", + "content": " denotes the continuous time step, and " + }, + { + "bbox": [ + 55, + 401, + 296, + 475 + ], + "type": "inline_equation", + "content": "v_{\\theta}" + }, + { + "bbox": [ + 55, + 401, + 296, + 475 + ], + "type": "text", + "content": " represents a neural network-parameterized velocity field. The condition " + }, + { + "bbox": [ + 55, + 401, + 296, + 475 + ], + "type": "inline_equation", + "content": "c" + }, + { + "bbox": [ + 55, + 401, + 296, + 475 + ], + "type": "text", + "content": " is incorporated for conditional flow implementations. Optimal transport flows enforce velocity alignment with the direction " + }, + { + "bbox": [ + 55, + 401, + 296, + 475 + ], + "type": "inline_equation", + "content": "(x_{1} - x_{0})" + }, + { + "bbox": [ + 55, + 401, + 296, + 475 + ], + "type": "text", + "content": " of the linear path pointing from " + }, + { + "bbox": [ + 55, + 401, + 296, + 475 + ], + "type": "inline_equation", + "content": "x_{0}" + }, + { + "bbox": [ + 55, + 401, + 296, + 475 + ], + "type": "text", + "content": " to " + }, + { + "bbox": [ + 55, + 401, + 296, + 475 + ], + "type": "inline_equation", + "content": "x_{1}" + }, + { + "bbox": [ + 55, + 401, + 296, + 475 + ], + "type": "text", + "content": " through regression:" + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 82, + 481, + 295, + 496 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 82, + 481, + 295, + 496 + ], + "spans": [ + { + "bbox": [ + 82, + 481, + 295, + 496 + ], + "type": "interline_equation", + "content": "\\mathcal {L} (\\theta) = \\mathbb {E} _ {t, x _ {0}, x _ {1}, c} \\| (x _ {1} - x _ {0}) - v _ {\\theta} (x _ {t}, t, c) \\| ^ {2}, \\tag {2}", + "image_path": "304215aa9da1cde6cbe344dbff36a76359c701993aec658a1fff3d5107153c56.jpg" + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 55, + 504, + 295, + 539 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 504, + 295, + 539 + ], + "spans": [ + { + "bbox": [ + 55, + 504, + 295, + 539 + ], + "type": "text", + "content": "where " + }, + { + "bbox": [ + 55, + 504, + 295, + 539 + ], + "type": "inline_equation", + "content": "x_{t} = (1 - t)x_{0} + tx_{1}" + }, + { + "bbox": [ + 55, + 504, + 295, + 539 + ], + "type": "text", + "content": " constitutes the linear interpolation between " + }, + { + "bbox": [ + 55, + 504, + 295, + 539 + ], + "type": "inline_equation", + "content": "x_{0}" + }, + { + "bbox": [ + 55, + 504, + 295, + 539 + ], + "type": "text", + "content": " and " + }, + { + "bbox": [ + 55, + 504, + 295, + 539 + ], + "type": "inline_equation", + "content": "x_{1}" + }, + { + "bbox": [ + 55, + 504, + 295, + 539 + ], + "type": "text", + "content": ". The learned velocity field enables target generation by solving the ODE in Equation 1." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 55, + 540, + 295, + 622 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 540, + 295, + 622 + ], + "spans": [ + { + "bbox": [ + 55, + 540, + 295, + 622 + ], + "type": "text", + "content": "Previous video-to-audio flow approaches [8, 43] (and related diffusion methods [25, 41]) use a Gaussian prior " + }, + { + "bbox": [ + 55, + 540, + 295, + 622 + ], + "type": "inline_equation", + "content": "p(x_0) = \\mathcal{N}(0,1)" + }, + { + "bbox": [ + 55, + 540, + 295, + 622 + ], + "type": "text", + "content": " for the source distribution, with audio as " + }, + { + "bbox": [ + 55, + 540, + 295, + 622 + ], + "type": "inline_equation", + "content": "p(x_{1})" + }, + { + "bbox": [ + 55, + 540, + 295, + 622 + ], + "type": "text", + "content": " and video as conditioning input " + }, + { + "bbox": [ + 55, + 540, + 295, + 622 + ], + "type": "inline_equation", + "content": "c" + }, + { + "bbox": [ + 55, + 540, + 295, + 622 + ], + "type": "text", + "content": ". During training, random noise " + }, + { + "bbox": [ + 55, + 540, + 295, + 622 + ], + "type": "inline_equation", + "content": "x_0" + }, + { + "bbox": [ + 55, + 540, + 295, + 622 + ], + "type": "text", + "content": " is paired with ground-truth audio " + }, + { + "bbox": [ + 55, + 540, + 295, + 622 + ], + "type": "inline_equation", + "content": "x_{1}" + }, + { + "bbox": [ + 55, + 540, + 295, + 622 + ], + "type": "text", + "content": ", prompting " + }, + { + "bbox": [ + 55, + 540, + 295, + 622 + ], + "type": "inline_equation", + "content": "v_{\\theta}" + }, + { + "bbox": [ + 55, + 540, + 295, + 622 + ], + "type": "text", + "content": " to model the paths from noise to audio under visual conditions." + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 55, + 624, + 296, + 659 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 624, + 296, + 659 + ], + "spans": [ + { + "bbox": [ + 55, + 624, + 296, + 659 + ], + "type": "text", + "content": "This work challenges the Gaussian prior by proposing a direct video-to-audio transport. Theoretically, FM permits arbitrary source distributions, motivating our formulation:" + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 132, + 668, + 295, + 682 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 132, + 668, + 295, + 682 + ], + "spans": [ + { + "bbox": [ + 132, + 668, + 295, + 682 + ], + "type": "interline_equation", + "content": "d x (t) = v _ {\\theta} \\left(x _ {t}, t\\right) d t, \\tag {3}", + "image_path": "cddedd00612ba49fab0148c1f27f728c53225b4966d19564cf8b6dc2f98c45ce.jpg" + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 55, + 689, + 296, + 715 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 689, + 296, + 715 + ], + "spans": [ + { + "bbox": [ + 55, + 689, + 296, + 715 + ], + "type": "text", + "content": "where " + }, + { + "bbox": [ + 55, + 689, + 296, + 715 + ], + "type": "inline_equation", + "content": "x_{t} = (1 - t)x_{v} + tx_{a}" + }, + { + "bbox": [ + 55, + 689, + 296, + 715 + ], + "type": "text", + "content": " linearly interpolates between paired video and audio samples " + }, + { + "bbox": [ + 55, + 689, + 296, + 715 + ], + "type": "inline_equation", + "content": "x_{v},x_{a}\\sim p(v,a)" + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 314, + 72, + 437, + 83 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 314, + 72, + 437, + 83 + ], + "spans": [ + { + "bbox": [ + 314, + 72, + 437, + 83 + ], + "type": "text", + "content": "3.2. VAFlow Architecture" + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 313, + 91, + 555, + 259 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 91, + 555, + 259 + ], + "spans": [ + { + "bbox": [ + 313, + 91, + 555, + 259 + ], + "type": "text", + "content": "Model Overview. As illustrated in Figure 2, the input video features " + }, + { + "bbox": [ + 313, + 91, + 555, + 259 + ], + "type": "inline_equation", + "content": "x_{v}" + }, + { + "bbox": [ + 313, + 91, + 555, + 259 + ], + "type": "text", + "content": " are extracted via a visual encoder, while the target audio waveform can be obtained from latent representations " + }, + { + "bbox": [ + 313, + 91, + 555, + 259 + ], + "type": "inline_equation", + "content": "x_{a}" + }, + { + "bbox": [ + 313, + 91, + 555, + 259 + ], + "type": "text", + "content": " through WaveVAE. VAFlow attempts to establish FM between the video feature space " + }, + { + "bbox": [ + 313, + 91, + 555, + 259 + ], + "type": "inline_equation", + "content": "x_{v}" + }, + { + "bbox": [ + 313, + 91, + 555, + 259 + ], + "type": "text", + "content": " and audio latent space " + }, + { + "bbox": [ + 313, + 91, + 555, + 259 + ], + "type": "inline_equation", + "content": "x_{a}" + }, + { + "bbox": [ + 313, + 91, + 555, + 259 + ], + "type": "text", + "content": ", enabling direct audio generation from video features via ODE solving instead of Gaussian noise initialization. However, since FM requires consistent dimensional shapes between source and target distributions (see Equation 3), the inherent spatiotemporal resolution mismatch between " + }, + { + "bbox": [ + 313, + 91, + 555, + 259 + ], + "type": "inline_equation", + "content": "x_{v}" + }, + { + "bbox": [ + 313, + 91, + 555, + 259 + ], + "type": "text", + "content": " and " + }, + { + "bbox": [ + 313, + 91, + 555, + 259 + ], + "type": "inline_equation", + "content": "x_{a}" + }, + { + "bbox": [ + 313, + 91, + 555, + 259 + ], + "type": "text", + "content": " necessitates shape alignment. To address this, we employ a 1D VAE (namely alignment VAE) that compresses interpolated " + }, + { + "bbox": [ + 313, + 91, + 555, + 259 + ], + "type": "inline_equation", + "content": "x_{v}" + }, + { + "bbox": [ + 313, + 91, + 555, + 259 + ], + "type": "text", + "content": " into aligned features " + }, + { + "bbox": [ + 313, + 91, + 555, + 259 + ], + "type": "inline_equation", + "content": "x_{v}^{\\mathrm{AE}}" + }, + { + "bbox": [ + 313, + 91, + 555, + 259 + ], + "type": "text", + "content": " matching " + }, + { + "bbox": [ + 313, + 91, + 555, + 259 + ], + "type": "inline_equation", + "content": "x_{a}" + }, + { + "bbox": [ + 313, + 91, + 555, + 259 + ], + "type": "text", + "content": "'s dimensions." + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 313, + 262, + 554, + 357 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 262, + 554, + 357 + ], + "spans": [ + { + "bbox": [ + 313, + 262, + 554, + 357 + ], + "type": "text", + "content": "A diffusion transformer (DiT) subsequently works as a velocity field estimator to capture the transport dynamics between " + }, + { + "bbox": [ + 313, + 262, + 554, + 357 + ], + "type": "inline_equation", + "content": "x_{v}^{\\mathrm{AE}}" + }, + { + "bbox": [ + 313, + 262, + 554, + 357 + ], + "type": "text", + "content": " and " + }, + { + "bbox": [ + 313, + 262, + 554, + 357 + ], + "type": "inline_equation", + "content": "x_{a}" + }, + { + "bbox": [ + 313, + 262, + 554, + 357 + ], + "type": "text", + "content": ". This enables continuous trajectory sampling from " + }, + { + "bbox": [ + 313, + 262, + 554, + 357 + ], + "type": "inline_equation", + "content": "x_{v}^{\\mathrm{AE}}" + }, + { + "bbox": [ + 313, + 262, + 554, + 357 + ], + "type": "text", + "content": " to " + }, + { + "bbox": [ + 313, + 262, + 554, + 357 + ], + "type": "inline_equation", + "content": "x_{a}" + }, + { + "bbox": [ + 313, + 262, + 554, + 357 + ], + "type": "text", + "content": " through ODE integration. The sampled latent " + }, + { + "bbox": [ + 313, + 262, + 554, + 357 + ], + "type": "inline_equation", + "content": "\\hat{x}_{a}" + }, + { + "bbox": [ + 313, + 262, + 554, + 357 + ], + "type": "text", + "content": " is finally decoded by WaveVAE to produce the output audio. The following contents detail: the visual/audio representations " + }, + { + "bbox": [ + 313, + 262, + 554, + 357 + ], + "type": "inline_equation", + "content": "(x_{v}, x_{v}^{\\mathrm{AE}}, x_{a})" + }, + { + "bbox": [ + 313, + 262, + 554, + 357 + ], + "type": "text", + "content": ", and the velocity field estimator in VAFlow." + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 313, + 359, + 555, + 418 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 359, + 555, + 418 + ], + "spans": [ + { + "bbox": [ + 313, + 359, + 555, + 418 + ], + "type": "text", + "content": "Video and Audio Representations. For an input video-audio pair " + }, + { + "bbox": [ + 313, + 359, + 555, + 418 + ], + "type": "inline_equation", + "content": "(v, a)" + }, + { + "bbox": [ + 313, + 359, + 555, + 418 + ], + "type": "text", + "content": ", the audio waveform " + }, + { + "bbox": [ + 313, + 359, + 555, + 418 + ], + "type": "inline_equation", + "content": "a" + }, + { + "bbox": [ + 313, + 359, + 555, + 418 + ], + "type": "text", + "content": " is compressed by a pretrained 1D WaveVAE's [11] encoder into " + }, + { + "bbox": [ + 313, + 359, + 555, + 418 + ], + "type": "inline_equation", + "content": "x_{a} \\in \\mathbb{R}^{T_{a} \\times D_{a}}" + }, + { + "bbox": [ + 313, + 359, + 555, + 418 + ], + "type": "text", + "content": ". This latent " + }, + { + "bbox": [ + 313, + 359, + 555, + 418 + ], + "type": "inline_equation", + "content": "x_{a}" + }, + { + "bbox": [ + 313, + 359, + 555, + 418 + ], + "type": "text", + "content": " serves as the target for flow generation and is later decoded to reconstruct the waveform." + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "spans": [ + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "text", + "content": "For video representation, a pretrained visual encoder encodes " + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "inline_equation", + "content": "v" + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "text", + "content": " into " + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "inline_equation", + "content": "x_v \\in \\mathbb{R}^{T_v \\times D_v}" + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "text", + "content": " (e.g., CLIP [32], CAVP [25], or VidTok [38] as encoder). Given the discrepancies (e.g., " + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "inline_equation", + "content": "x_a \\in \\mathbb{R}^{215 \\times 64}" + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "text", + "content": " versus " + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "inline_equation", + "content": "x_v \\in \\mathbb{R}^{100 \\times 768}" + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "text", + "content": " for a 10-second clip), we propose an alignment VAE " + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "inline_equation", + "content": "\\epsilon_{\\phi}" + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "text", + "content": " to further compress and align video latents with audio latents' space while preserving essential visual information. Specifically, we utilize a 1D convolutional encoder-decoder as " + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "inline_equation", + "content": "\\epsilon_{\\phi}" + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "text", + "content": " to interpolate and project " + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "inline_equation", + "content": "x_v" + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "text", + "content": " into a variational space " + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "inline_equation", + "content": "x_v^{\\mathrm{AE}} \\in \\mathbb{R}^{T_a \\times D_a}" + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "text", + "content": ", matching the shape of " + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "inline_equation", + "content": "x_a" + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "text", + "content": ". The encoder of " + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "inline_equation", + "content": "\\epsilon_{\\phi}" + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "text", + "content": " predicts the parameters " + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "inline_equation", + "content": "\\mu_{x_v^{\\mathrm{AE}}}, \\sigma_{x_v^{\\mathrm{AE}}} = E_{\\epsilon_{\\phi}}(x_v)" + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "text", + "content": " for a Gaussian distribution, from which " + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "inline_equation", + "content": "x_v^{\\mathrm{AE}} \\sim \\mathcal{N}(\\mu_{x_v^{\\mathrm{AE}}}, \\sigma_{x_v^{\\mathrm{AE}}})" + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "text", + "content": " is sampled. The decoder reconstructs " + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "inline_equation", + "content": "x_v" + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "text", + "content": " as " + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "inline_equation", + "content": "\\hat{x}_v = D_{\\epsilon_{\\phi}}(x_v^{\\mathrm{AE}})" + }, + { + "bbox": [ + 313, + 422, + 555, + 616 + ], + "type": "text", + "content": ". This process ensures dimensional alignment with audio latents while retaining critical spatiotemporal features of the original video representation." + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 313, + 617, + 556, + 714 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 617, + 556, + 714 + ], + "spans": [ + { + "bbox": [ + 313, + 617, + 556, + 714 + ], + "type": "text", + "content": "Velocity Estimator. Built upon the diffusion transformer (DiT) architecture [11], the velocity estimator processes two key inputs: time embeddings of " + }, + { + "bbox": [ + 313, + 617, + 556, + 714 + ], + "type": "inline_equation", + "content": "t" + }, + { + "bbox": [ + 313, + 617, + 556, + 714 + ], + "type": "text", + "content": " and the interpolated state " + }, + { + "bbox": [ + 313, + 617, + 556, + 714 + ], + "type": "inline_equation", + "content": "x_{t} = (1 - t)x_{v}^{\\mathrm{AE}} + tx_{a}" + }, + { + "bbox": [ + 313, + 617, + 556, + 714 + ], + "type": "text", + "content": ", following the Equation 3. The time embedding is concatenated with the audio sequence to form an input tensor " + }, + { + "bbox": [ + 313, + 617, + 556, + 714 + ], + "type": "inline_equation", + "content": "\\in \\mathbb{R}^{(T_a + 1)\\times D_a}" + }, + { + "bbox": [ + 313, + 617, + 556, + 714 + ], + "type": "text", + "content": ", which is processed through stacked transformer blocks. The first " + }, + { + "bbox": [ + 313, + 617, + 556, + 714 + ], + "type": "inline_equation", + "content": "T_{a}" + }, + { + "bbox": [ + 313, + 617, + 556, + 714 + ], + "type": "text", + "content": " dimensions of the output sequence constitute the pre" + } + ] + } + ], + "index": 17 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 293, + 748, + 318, + 758 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 293, + 748, + 318, + 758 + ], + "spans": [ + { + "bbox": [ + 293, + 748, + 318, + 758 + ], + "type": "text", + "content": "11779" + } + ] + } + ], + "index": 18 + } + ], + "page_size": [ + 612, + 792 + ], + "page_idx": 2 + }, + { + "para_blocks": [ + { + "type": "image", + "bbox": [ + 59, + 71, + 550, + 178 + ], + "blocks": [ + { + "bbox": [ + 59, + 71, + 550, + 178 + ], + "lines": [ + { + "bbox": [ + 59, + 71, + 550, + 178 + ], + "spans": [ + { + "bbox": [ + 59, + 71, + 550, + 178 + ], + "type": "image", + "image_path": "ff1ff64fb37921e45705859648878aee89608f3208a30432c279d825228be95d.jpg" + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 55, + 190, + 555, + 257 + ], + "lines": [ + { + "bbox": [ + 55, + 190, + 555, + 257 + ], + "spans": [ + { + "bbox": [ + 55, + 190, + 555, + 257 + ], + "type": "text", + "content": "Figure 2. VAFlow Architecture. The figure illustrates VAFow in the inference manner. A visual encoder initially extracts temporal features " + }, + { + "bbox": [ + 55, + 190, + 555, + 257 + ], + "type": "inline_equation", + "content": "x_{v} \\in \\mathbb{R}^{T_{v} \\times D_{v}}" + }, + { + "bbox": [ + 55, + 190, + 555, + 257 + ], + "type": "text", + "content": " from variable-frame-rate video inputs. Then, a 1D alignment VAE (Variational Autoencoder) " + }, + { + "bbox": [ + 55, + 190, + 555, + 257 + ], + "type": "inline_equation", + "content": "\\epsilon_{\\phi}" + }, + { + "bbox": [ + 55, + 190, + 555, + 257 + ], + "type": "text", + "content": " compresses interpolated " + }, + { + "bbox": [ + 55, + 190, + 555, + 257 + ], + "type": "inline_equation", + "content": "x_{v}" + }, + { + "bbox": [ + 55, + 190, + 555, + 257 + ], + "type": "text", + "content": " into audio-aligned latent features " + }, + { + "bbox": [ + 55, + 190, + 555, + 257 + ], + "type": "inline_equation", + "content": "x_{v}^{\\mathrm{AE}}" + }, + { + "bbox": [ + 55, + 190, + 555, + 257 + ], + "type": "text", + "content": " with matched dimensions " + }, + { + "bbox": [ + 55, + 190, + 555, + 257 + ], + "type": "inline_equation", + "content": "(T_{v}^{\\mathrm{AE}} \\times D_{v}^{\\mathrm{AE}} = T_{a} \\times D_{a})" + }, + { + "bbox": [ + 55, + 190, + 555, + 257 + ], + "type": "text", + "content": ". Next, a diffusion transformer " + }, + { + "bbox": [ + 55, + 190, + 555, + 257 + ], + "type": "inline_equation", + "content": "(v_{\\theta})" + }, + { + "bbox": [ + 55, + 190, + 555, + 257 + ], + "type": "text", + "content": " predicts the velocity field for the transport " + }, + { + "bbox": [ + 55, + 190, + 555, + 257 + ], + "type": "inline_equation", + "content": "(x_{a} - x_{v}^{\\mathrm{AE}})" + }, + { + "bbox": [ + 55, + 190, + 555, + 257 + ], + "type": "text", + "content": ", and by solving the corresponding ODE (see Equation 3), an estimated " + }, + { + "bbox": [ + 55, + 190, + 555, + 257 + ], + "type": "inline_equation", + "content": "\\hat{x}_{a}" + }, + { + "bbox": [ + 55, + 190, + 555, + 257 + ], + "type": "text", + "content": " is obtained. Finally, the WaveVAE decoder converts " + }, + { + "bbox": [ + 55, + 190, + 555, + 257 + ], + "type": "inline_equation", + "content": "\\hat{x}_{a}" + }, + { + "bbox": [ + 55, + 190, + 555, + 257 + ], + "type": "text", + "content": " into the waveform representing the predicted audio. Note that during diffusion transformer prediction, " + }, + { + "bbox": [ + 55, + 190, + 555, + 257 + ], + "type": "inline_equation", + "content": "x_{v}" + }, + { + "bbox": [ + 55, + 190, + 555, + 257 + ], + "type": "text", + "content": " can optionally be incorporated via cross-attention, detail discussion is conducted in the experiment Section 4.5." + } + ] + } + ], + "index": 1, + "angle": 0, + "type": "image_caption" + } + ], + "index": 0 + }, + { + "bbox": [ + 54, + 277, + 294, + 397 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 54, + 277, + 294, + 397 + ], + "spans": [ + { + "bbox": [ + 54, + 277, + 294, + 397 + ], + "type": "text", + "content": "dicted velocity " + }, + { + "bbox": [ + 54, + 277, + 294, + 397 + ], + "type": "inline_equation", + "content": "v_{\\theta}(x_t, t)" + }, + { + "bbox": [ + 54, + 277, + 294, + 397 + ], + "type": "text", + "content": ", with 1D Rotary Position Embeddings (RoPE [36]) to preserve positional awareness in attention layers. Notably, while the interpolated state " + }, + { + "bbox": [ + 54, + 277, + 294, + 397 + ], + "type": "inline_equation", + "content": "x_t" + }, + { + "bbox": [ + 54, + 277, + 294, + 397 + ], + "type": "text", + "content": " implicitly encodes visual information through " + }, + { + "bbox": [ + 54, + 277, + 294, + 397 + ], + "type": "inline_equation", + "content": "x_v^{\\mathrm{AE}}" + }, + { + "bbox": [ + 54, + 277, + 294, + 397 + ], + "type": "text", + "content": ", this signal diminishes as " + }, + { + "bbox": [ + 54, + 277, + 294, + 397 + ], + "type": "inline_equation", + "content": "t \\to 1" + }, + { + "bbox": [ + 54, + 277, + 294, + 397 + ], + "type": "text", + "content": ". To address this limitation, we retain cross-attention layers in the transformer blocks that persistently integrate the original video features " + }, + { + "bbox": [ + 54, + 277, + 294, + 397 + ], + "type": "inline_equation", + "content": "x_v" + }, + { + "bbox": [ + 54, + 277, + 294, + 397 + ], + "type": "text", + "content": ". Further studies (Section 4.5) demonstrate that this architectural choice is critical to enhance visual perception throughout the flow trajectory and improve audio generation quality." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 55, + 404, + 228, + 417 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 404, + 228, + 417 + ], + "spans": [ + { + "bbox": [ + 55, + 404, + 228, + 417 + ], + "type": "text", + "content": "3.3. VAFlow Training and Inference" + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 55, + 422, + 295, + 470 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 422, + 295, + 470 + ], + "spans": [ + { + "bbox": [ + 55, + 422, + 295, + 470 + ], + "type": "text", + "content": "We conduct three-stage training: (1) pretrain the alignment VAE, (2) train the VAFlow estimator with frozen alignment VAE, and (3) jointly fine-tune both alignment VAE and estimator. We detail training and inference as follows:" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 55, + 470, + 295, + 544 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 470, + 295, + 544 + ], + "spans": [ + { + "bbox": [ + 55, + 470, + 295, + 544 + ], + "type": "text", + "content": "Alignment VAE Training. Given encoded visual features " + }, + { + "bbox": [ + 55, + 470, + 295, + 544 + ], + "type": "inline_equation", + "content": "x_{v} \\in \\mathbb{R}^{T_{v} \\times D_{v}}" + }, + { + "bbox": [ + 55, + 470, + 295, + 544 + ], + "type": "text", + "content": ", the alignment VAE encoder predicts distribution parameters " + }, + { + "bbox": [ + 55, + 470, + 295, + 544 + ], + "type": "inline_equation", + "content": "(\\mu_{x_{v}^{\\mathrm{AE}}}, \\sigma_{x_{v}^{\\mathrm{AE}}})" + }, + { + "bbox": [ + 55, + 470, + 295, + 544 + ], + "type": "text", + "content": ", while the decoder reconstructs " + }, + { + "bbox": [ + 55, + 470, + 295, + 544 + ], + "type": "inline_equation", + "content": "\\hat{x}_{v}" + }, + { + "bbox": [ + 55, + 470, + 295, + 544 + ], + "type": "text", + "content": " from a sampled " + }, + { + "bbox": [ + 55, + 470, + 295, + 544 + ], + "type": "inline_equation", + "content": "x_{v}^{\\mathrm{AE}} \\sim \\mathcal{N}(\\mu_{x_{v}^{\\mathrm{AE}}}, \\sigma_{x_{v}^{\\mathrm{AE}}})" + }, + { + "bbox": [ + 55, + 470, + 295, + 544 + ], + "type": "text", + "content": ". The model optimizes three objectives to preserve visual semantics while regularizing " + }, + { + "bbox": [ + 55, + 470, + 295, + 544 + ], + "type": "inline_equation", + "content": "x_{v}^{\\mathrm{AE}}" + }, + { + "bbox": [ + 55, + 470, + 295, + 544 + ], + "type": "text", + "content": " within " + }, + { + "bbox": [ + 55, + 470, + 295, + 544 + ], + "type": "inline_equation", + "content": "\\mathbb{R}^{T_{a} \\times D_{a}}" + }, + { + "bbox": [ + 55, + 470, + 295, + 544 + ], + "type": "text", + "content": " Gaussian space:" + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 55, + 545, + 288, + 556 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 545, + 288, + 556 + ], + "spans": [ + { + "bbox": [ + 55, + 545, + 288, + 556 + ], + "type": "text", + "content": "- KL Regularization: enforces Gaussian latent priors via:" + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 91, + 566, + 295, + 580 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 91, + 566, + 295, + 580 + ], + "spans": [ + { + "bbox": [ + 91, + 566, + 295, + 580 + ], + "type": "interline_equation", + "content": "\\mathcal {L} _ {\\mathrm {K L}} = D _ {\\mathrm {K L}} \\left(\\mathcal {N} \\left(\\mu_ {x _ {0} ^ {\\mathrm {A E}}}, \\sigma_ {x _ {0} ^ {\\mathrm {A E}}}\\right) \\| \\mathcal {N} (0, I)\\right). \\tag {4}", + "image_path": "f14f9e7bb3df9bd43d09feba44d2bd6f84d4ea7680838ee5ccb86b74bf38a6d3.jpg" + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 55, + 588, + 288, + 600 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 588, + 288, + 600 + ], + "spans": [ + { + "bbox": [ + 55, + 588, + 288, + 600 + ], + "type": "text", + "content": "- Reconstruction Loss: ensures feature-level consistency:" + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 137, + 609, + 295, + 622 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 137, + 609, + 295, + 622 + ], + "spans": [ + { + "bbox": [ + 137, + 609, + 295, + 622 + ], + "type": "interline_equation", + "content": "\\mathcal {L} _ {\\text {r e c}} = \\left\\| x _ {v} - \\hat {x} _ {v} \\right\\| _ {2} ^ {2}. \\tag {5}", + "image_path": "15acbf16995dfcb5874d4441483c99dd90ab066f68d9c081d36b6a9cdada5bd5.jpg" + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 55, + 632, + 293, + 643 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 632, + 293, + 643 + ], + "spans": [ + { + "bbox": [ + 55, + 632, + 293, + 643 + ], + "type": "text", + "content": "- Contrastive Alignment: strengthens feature preservation:" + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 70, + 651, + 295, + 680 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 70, + 651, + 295, + 680 + ], + "spans": [ + { + "bbox": [ + 70, + 651, + 295, + 680 + ], + "type": "interline_equation", + "content": "\\mathcal {L} _ {\\text {c o n}} = - \\sum_ {s \\in \\mathcal {S}} \\log \\frac {\\exp \\left(f _ {s} \\left(\\hat {x} _ {v}\\right) \\cdot f _ {s} \\left(x _ {v}\\right) / \\tau\\right)}{\\sum_ {x ^ {\\prime} \\in \\mathcal {B}} \\exp \\left(f _ {s} \\left(\\hat {x} _ {v}\\right) \\cdot f _ {s} \\left(x ^ {\\prime}\\right) / \\tau\\right)}, \\tag {6}", + "image_path": "b8dff3acf95149d88232fb483c3d57074962431fbb34e62bd27f8b1fdbcc3182.jpg" + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 63, + 689, + 295, + 713 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 63, + 689, + 295, + 713 + ], + "spans": [ + { + "bbox": [ + 63, + 689, + 295, + 713 + ], + "type": "text", + "content": "where " + }, + { + "bbox": [ + 63, + 689, + 295, + 713 + ], + "type": "inline_equation", + "content": "S = \\{\\mathrm{frame},\\mathrm{clip},\\mathrm{global}\\}" + }, + { + "bbox": [ + 63, + 689, + 295, + 713 + ], + "type": "text", + "content": " defines temporal granularities and " + }, + { + "bbox": [ + 63, + 689, + 295, + 713 + ], + "type": "inline_equation", + "content": "\\tau = 0.7" + }, + { + "bbox": [ + 63, + 689, + 295, + 713 + ], + "type": "text", + "content": " serves as a temperature parameter [6]." + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 321, + 277, + 555, + 338 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 321, + 277, + 555, + 338 + ], + "spans": [ + { + "bbox": [ + 321, + 277, + 555, + 338 + ], + "type": "text", + "content": "For scale " + }, + { + "bbox": [ + 321, + 277, + 555, + 338 + ], + "type": "inline_equation", + "content": "s" + }, + { + "bbox": [ + 321, + 277, + 555, + 338 + ], + "type": "text", + "content": ", " + }, + { + "bbox": [ + 321, + 277, + 555, + 338 + ], + "type": "inline_equation", + "content": "f_{s}(\\cdot)" + }, + { + "bbox": [ + 321, + 277, + 555, + 338 + ], + "type": "text", + "content": " extracts features via three ways: frame-level: single-frame sampling + mean pooling; clip-level: " + }, + { + "bbox": [ + 321, + 277, + 555, + 338 + ], + "type": "inline_equation", + "content": "K = 8" + }, + { + "bbox": [ + 321, + 277, + 555, + 338 + ], + "type": "text", + "content": "-adjacent-frame sampling + pooling; global-level: full temporal mean pooling at " + }, + { + "bbox": [ + 321, + 277, + 555, + 338 + ], + "type": "inline_equation", + "content": "T_{v}" + }, + { + "bbox": [ + 321, + 277, + 555, + 338 + ], + "type": "text", + "content": " dimension of feature " + }, + { + "bbox": [ + 321, + 277, + 555, + 338 + ], + "type": "inline_equation", + "content": "x" + }, + { + "bbox": [ + 321, + 277, + 555, + 338 + ], + "type": "text", + "content": ". Negative samples " + }, + { + "bbox": [ + 321, + 277, + 555, + 338 + ], + "type": "inline_equation", + "content": "x'" + }, + { + "bbox": [ + 321, + 277, + 555, + 338 + ], + "type": "text", + "content": " are drawn from training batch " + }, + { + "bbox": [ + 321, + 277, + 555, + 338 + ], + "type": "inline_equation", + "content": "\\mathcal{B}" + }, + { + "bbox": [ + 321, + 277, + 555, + 338 + ], + "type": "text", + "content": "." + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 314, + 339, + 506, + 351 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 314, + 339, + 506, + 351 + ], + "spans": [ + { + "bbox": [ + 314, + 339, + 506, + 351 + ], + "type": "text", + "content": "The composite objective integrates these terms:" + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 353, + 359, + 553, + 371 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 353, + 359, + 553, + 371 + ], + "spans": [ + { + "bbox": [ + 353, + 359, + 553, + 371 + ], + "type": "interline_equation", + "content": "\\mathcal {L} _ {\\mathrm {V A E}} = \\lambda_ {\\mathrm {K L}} \\mathcal {L} _ {\\mathrm {K L}} + \\lambda_ {\\mathrm {r e c}} \\mathcal {L} _ {\\mathrm {r e c}} + \\lambda_ {\\mathrm {c o n}} \\mathcal {L} _ {\\mathrm {c o n}}, \\tag {7}", + "image_path": "623d06976355c95defdc32e63705c5584c9fcffa8b59e1f0084d01fea5f2d049.jpg" + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 313, + 379, + 514, + 391 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 379, + 514, + 391 + ], + "spans": [ + { + "bbox": [ + 313, + 379, + 514, + 391 + ], + "type": "text", + "content": "with " + }, + { + "bbox": [ + 313, + 379, + 514, + 391 + ], + "type": "inline_equation", + "content": "\\lambda_{\\mathrm{KL}} = 8\\times 10^{-4}" + }, + { + "bbox": [ + 313, + 379, + 514, + 391 + ], + "type": "inline_equation", + "content": "\\lambda_{\\mathrm{rec}} = 1.0" + }, + { + "bbox": [ + 313, + 379, + 514, + 391 + ], + "type": "text", + "content": " , and " + }, + { + "bbox": [ + 313, + 379, + 514, + 391 + ], + "type": "inline_equation", + "content": "\\lambda_{\\mathrm{con}} = 1.0" + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 313, + 392, + 554, + 427 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 392, + 554, + 427 + ], + "spans": [ + { + "bbox": [ + 313, + 392, + 554, + 427 + ], + "type": "text", + "content": "Velocity Estimator Training Given encoded features " + }, + { + "bbox": [ + 313, + 392, + 554, + 427 + ], + "type": "inline_equation", + "content": "x_{v}" + }, + { + "bbox": [ + 313, + 392, + 554, + 427 + ], + "type": "text", + "content": ", " + }, + { + "bbox": [ + 313, + 392, + 554, + 427 + ], + "type": "inline_equation", + "content": "x_{v}^{\\mathrm{AE}}" + }, + { + "bbox": [ + 313, + 392, + 554, + 427 + ], + "type": "text", + "content": ", and " + }, + { + "bbox": [ + 313, + 392, + 554, + 427 + ], + "type": "inline_equation", + "content": "x_{a}" + }, + { + "bbox": [ + 313, + 392, + 554, + 427 + ], + "type": "text", + "content": ", we train the velocity estimator " + }, + { + "bbox": [ + 313, + 392, + 554, + 427 + ], + "type": "inline_equation", + "content": "v_{\\theta}" + }, + { + "bbox": [ + 313, + 392, + 554, + 427 + ], + "type": "text", + "content": " to predict the linear transport direction " + }, + { + "bbox": [ + 313, + 392, + 554, + 427 + ], + "type": "inline_equation", + "content": "x_{a} - x_{v}^{\\mathrm{AE}}" + }, + { + "bbox": [ + 313, + 392, + 554, + 427 + ], + "type": "text", + "content": " through:" + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 325, + 434, + 553, + 452 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 325, + 434, + 553, + 452 + ], + "spans": [ + { + "bbox": [ + 325, + 434, + 553, + 452 + ], + "type": "interline_equation", + "content": "\\mathcal {L} (\\theta) = \\mathbb {E} _ {t \\sim \\mathcal {U} (0, 1)} \\left\\| \\left(x _ {a} - x _ {v} ^ {\\mathrm {A E}}\\right) - v _ {\\theta} \\left(x _ {t}, t, x _ {v}\\right) \\right\\| _ {2} ^ {2}, \\tag {8}", + "image_path": "bc5b7c0f3a13f909dd14ae8f9d7c3e73c47c9708cc3f2136b7d534407bce6def.jpg" + } + ] + } + ], + "index": 18 + }, + { + "bbox": [ + 313, + 458, + 554, + 518 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 458, + 554, + 518 + ], + "spans": [ + { + "bbox": [ + 313, + 458, + 554, + 518 + ], + "type": "text", + "content": "where " + }, + { + "bbox": [ + 313, + 458, + 554, + 518 + ], + "type": "inline_equation", + "content": "x_{t} = (1 - t)x_{v}^{\\mathrm{AE}} + tx_{a}" + }, + { + "bbox": [ + 313, + 458, + 554, + 518 + ], + "type": "text", + "content": ". For the conditioned version, visual features " + }, + { + "bbox": [ + 313, + 458, + 554, + 518 + ], + "type": "inline_equation", + "content": "x_{v}" + }, + { + "bbox": [ + 313, + 458, + 554, + 518 + ], + "type": "text", + "content": " are incorporated via cross-attention (with a " + }, + { + "bbox": [ + 313, + 458, + 554, + 518 + ], + "type": "inline_equation", + "content": "10\\%" + }, + { + "bbox": [ + 313, + 458, + 554, + 518 + ], + "type": "text", + "content": " probability of zeroing out " + }, + { + "bbox": [ + 313, + 458, + 554, + 518 + ], + "type": "inline_equation", + "content": "x_{v}" + }, + { + "bbox": [ + 313, + 458, + 554, + 518 + ], + "type": "text", + "content": " during training). The vanilla version omits this conditioning, simplifying the input of estimator in Equation 8 to " + }, + { + "bbox": [ + 313, + 458, + 554, + 518 + ], + "type": "inline_equation", + "content": "v_{\\theta}(x_t,t)" + }, + { + "bbox": [ + 313, + 458, + 554, + 518 + ], + "type": "text", + "content": "." + } + ] + } + ], + "index": 19 + }, + { + "bbox": [ + 313, + 519, + 555, + 638 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 519, + 555, + 638 + ], + "spans": [ + { + "bbox": [ + 313, + 519, + 555, + 638 + ], + "type": "text", + "content": "Joint Turning After separate training of the alignment VAE and the velocity estimator, a joint optimization stage is performed. In this stage, the VAE encoder and " + }, + { + "bbox": [ + 313, + 519, + 555, + 638 + ], + "type": "inline_equation", + "content": "v_{\\theta}" + }, + { + "bbox": [ + 313, + 519, + 555, + 638 + ], + "type": "text", + "content": " are co-trained (discarding the VAE decoder), while the WaveVAE and visual feature extractor remain frozen. For a pair " + }, + { + "bbox": [ + 313, + 519, + 555, + 638 + ], + "type": "inline_equation", + "content": "(v, a)" + }, + { + "bbox": [ + 313, + 519, + 555, + 638 + ], + "type": "text", + "content": ", the visual extractor and WaveVAE produce " + }, + { + "bbox": [ + 313, + 519, + 555, + 638 + ], + "type": "inline_equation", + "content": "x_{v}" + }, + { + "bbox": [ + 313, + 519, + 555, + 638 + ], + "type": "text", + "content": " and " + }, + { + "bbox": [ + 313, + 519, + 555, + 638 + ], + "type": "inline_equation", + "content": "x_{a}" + }, + { + "bbox": [ + 313, + 519, + 555, + 638 + ], + "type": "text", + "content": ". The VAE encoder compresses " + }, + { + "bbox": [ + 313, + 519, + 555, + 638 + ], + "type": "inline_equation", + "content": "x_{v}" + }, + { + "bbox": [ + 313, + 519, + 555, + 638 + ], + "type": "text", + "content": " to obtain " + }, + { + "bbox": [ + 313, + 519, + 555, + 638 + ], + "type": "inline_equation", + "content": "x_{v}^{\\mathrm{AE}} \\sim q_{\\phi}(x_{v}^{\\mathrm{AE}}|x_{v})" + }, + { + "bbox": [ + 313, + 519, + 555, + 638 + ], + "type": "text", + "content": " which together with " + }, + { + "bbox": [ + 313, + 519, + 555, + 638 + ], + "type": "inline_equation", + "content": "x_{a}" + }, + { + "bbox": [ + 313, + 519, + 555, + 638 + ], + "type": "text", + "content": " and a random " + }, + { + "bbox": [ + 313, + 519, + 555, + 638 + ], + "type": "inline_equation", + "content": "t \\sim \\mathcal{U}(0,1)" + }, + { + "bbox": [ + 313, + 519, + 555, + 638 + ], + "type": "text", + "content": " forms " + }, + { + "bbox": [ + 313, + 519, + 555, + 638 + ], + "type": "inline_equation", + "content": "x_{t}" + }, + { + "bbox": [ + 313, + 519, + 555, + 638 + ], + "type": "text", + "content": ". The joint objective integrates KL regularization in Equation 4 with velocity prediction in Equation 8:" + } + ] + } + ], + "index": 20 + }, + { + "bbox": [ + 314, + 644, + 555, + 670 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 314, + 644, + 555, + 670 + ], + "spans": [ + { + "bbox": [ + 314, + 644, + 555, + 670 + ], + "type": "interline_equation", + "content": "\\mathcal {L} (\\theta , \\phi_ {\\text {E n c}}) = \\lambda_ {\\mathrm {K L}} \\mathcal {L} _ {\\mathrm {K L}} + \\mathbb {E} \\| \\left(x _ {a} - x _ {v} ^ {\\mathrm {A E}}\\right) - v _ {\\theta} \\left(x _ {t}, t, x _ {v}\\right) \\| _ {2} ^ {2}. \\tag {9}", + "image_path": "280758403897054e3bb2d7c18b999ddf22746ceffdcadb3988d9e62bb1ee9931.jpg" + } + ] + } + ], + "index": 21 + }, + { + "bbox": [ + 313, + 677, + 555, + 713 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 677, + 555, + 713 + ], + "spans": [ + { + "bbox": [ + 313, + 677, + 555, + 713 + ], + "type": "text", + "content": "Inference Upon training, VAFlow synthesizes audio " + }, + { + "bbox": [ + 313, + 677, + 555, + 713 + ], + "type": "inline_equation", + "content": "a" + }, + { + "bbox": [ + 313, + 677, + 555, + 713 + ], + "type": "text", + "content": " from an input video " + }, + { + "bbox": [ + 313, + 677, + 555, + 713 + ], + "type": "inline_equation", + "content": "v" + }, + { + "bbox": [ + 313, + 677, + 555, + 713 + ], + "type": "text", + "content": " through the following pipeline: The visual extractor first encodes " + }, + { + "bbox": [ + 313, + 677, + 555, + 713 + ], + "type": "inline_equation", + "content": "v" + }, + { + "bbox": [ + 313, + 677, + 555, + 713 + ], + "type": "text", + "content": " into a feature " + }, + { + "bbox": [ + 313, + 677, + 555, + 713 + ], + "type": "inline_equation", + "content": "x_{v}" + }, + { + "bbox": [ + 313, + 677, + 555, + 713 + ], + "type": "text", + "content": ", which is" + } + ] + } + ], + "index": 22 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "spans": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "type": "text", + "content": "11780" + } + ] + } + ], + "index": 23 + } + ], + "page_size": [ + 612, + 792 + ], + "page_idx": 3 + }, + { + "para_blocks": [ + { + "bbox": [ + 55, + 72, + 294, + 170 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 72, + 294, + 170 + ], + "spans": [ + { + "bbox": [ + 55, + 72, + 294, + 170 + ], + "type": "text", + "content": "compressed by the alignment VAE encoder into distribution parameters " + }, + { + "bbox": [ + 55, + 72, + 294, + 170 + ], + "type": "inline_equation", + "content": "(\\mu_{x_v^{\\mathrm{AE}}},\\sigma_{x_v^{\\mathrm{AE}}})" + }, + { + "bbox": [ + 55, + 72, + 294, + 170 + ], + "type": "text", + "content": ". A latent sample " + }, + { + "bbox": [ + 55, + 72, + 294, + 170 + ], + "type": "inline_equation", + "content": "x_{v}^{\\mathrm{AE}}\\sim \\mathcal{N}(\\mu_{x_v^{\\mathrm{AE}}},\\sigma_{x_v^{\\mathrm{AE}}}^2)" + }, + { + "bbox": [ + 55, + 72, + 294, + 170 + ], + "type": "text", + "content": " is drawn as the initial state " + }, + { + "bbox": [ + 55, + 72, + 294, + 170 + ], + "type": "inline_equation", + "content": "(x_0)" + }, + { + "bbox": [ + 55, + 72, + 294, + 170 + ], + "type": "text", + "content": " for solving the ODE in Equation 1 with learned estimator. The predicted " + }, + { + "bbox": [ + 55, + 72, + 294, + 170 + ], + "type": "inline_equation", + "content": "\\hat{x}_a(x_1)" + }, + { + "bbox": [ + 55, + 72, + 294, + 170 + ], + "type": "text", + "content": " is then decoded into waveform via the frozen WaveVAE. During inference, the variational sampling of " + }, + { + "bbox": [ + 55, + 72, + 294, + 170 + ], + "type": "inline_equation", + "content": "x_{v}^{\\mathrm{AE}}" + }, + { + "bbox": [ + 55, + 72, + 294, + 170 + ], + "type": "text", + "content": " introduces stochasticity - even with fixed ODE integration paths, distinct samples yield diverse outputs." + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 55, + 171, + 294, + 206 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 171, + 294, + 206 + ], + "spans": [ + { + "bbox": [ + 55, + 171, + 294, + 206 + ], + "type": "text", + "content": "For conditional velocity estimators, classifier-free guidance (CFG) [14] enhances synthesis quality by interpolating conditioned and unconditioned velocity predictions:" + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 63, + 213, + 294, + 239 + ], + "type": "interline_equation", + "angle": 0, + "lines": [ + { + "bbox": [ + 63, + 213, + 294, + 239 + ], + "spans": [ + { + "bbox": [ + 63, + 213, + 294, + 239 + ], + "type": "interline_equation", + "content": "v _ {\\theta} ^ {\\mathrm {C F G}} \\left(x _ {t}, t, x _ {v}\\right) = (1 + \\gamma) v _ {\\theta} \\left(x _ {t}, t, x _ {v}\\right) - \\gamma v _ {\\theta} \\left(x _ {t}, t, \\emptyset\\right), \\tag {10}", + "image_path": "a4867d985b258f515e4c44a33ae6656538e306923a15d020552a9ec42ff7ba35.jpg" + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 55, + 247, + 294, + 272 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 247, + 294, + 272 + ], + "spans": [ + { + "bbox": [ + 55, + 247, + 294, + 272 + ], + "type": "text", + "content": "where " + }, + { + "bbox": [ + 55, + 247, + 294, + 272 + ], + "type": "inline_equation", + "content": "\\gamma" + }, + { + "bbox": [ + 55, + 247, + 294, + 272 + ], + "type": "text", + "content": " controls guidance strength and " + }, + { + "bbox": [ + 55, + 247, + 294, + 272 + ], + "type": "inline_equation", + "content": "\\varnothing" + }, + { + "bbox": [ + 55, + 247, + 294, + 272 + ], + "type": "text", + "content": " denotes null (zero) conditioning." + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 55, + 282, + 136, + 295 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 282, + 136, + 295 + ], + "spans": [ + { + "bbox": [ + 55, + 282, + 136, + 295 + ], + "type": "text", + "content": "4. Experiments" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 55, + 301, + 171, + 314 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 301, + 171, + 314 + ], + "spans": [ + { + "bbox": [ + 55, + 301, + 171, + 314 + ], + "type": "text", + "content": "4.1. Experimental Setup" + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 55, + 319, + 294, + 450 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 319, + 294, + 450 + ], + "spans": [ + { + "bbox": [ + 55, + 319, + 294, + 450 + ], + "type": "text", + "content": "Datasets. We employ two datasets: VGGSound [5] and AudioSet-V2A [25]. For VGGSound, we adhere to the original split, which comprises 186k training samples and 15k testing samples. AudioSet-V2A contains 390k videos that are exclusively used for training. During training, the alignment VAE is trained on a combined set of VGGSound training and AudioSet-V2A to yield more generalized visual feature compression. The subsequent velocity estimator training and joint tuning are conducted solely on the VGGSound training set. At inference, we evaluate on the VGGSound test set as in prior work." + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 55, + 451, + 295, + 713 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 451, + 295, + 713 + ], + "spans": [ + { + "bbox": [ + 55, + 451, + 295, + 713 + ], + "type": "text", + "content": "Implementation Details. For the estimator's DiT architecture, we adapted the DiT implementation from StableAudio-Open-1.0 [11] (SDA1.0) by eliminating duration encoding, applying global time encoding solely to time steps, and restricting the cross-attention layer's key/value inputs to video features. For the visual extractor, we use three variants: CLIP [32] (a widely adopted visual semantic extractor in V2A tasks), CAVP [25] (which captures both visual semantic and temporal features), and VidTok [38] (a visual tokenizer for reconstruction). We train three alignment VAEs with comparable parameter scales—one for each visual extractor. In the velocity estimator training stage, the alignment VAE remains frozen; in the joint training stage, both the estimator and the alignment VAE are trained. In all stages, the audio WaveVAE and the visual extractor are kept frozen. Notably, we do not perform any further rectified iterations on the estimator; all VAFlow variants are trained for a single rectified iteration (Rectified-1, RF-1). During sampling, we employ ODE solvers of various orders—first-order Euler, second-order Midpoint, and fifth-order dopri5—with differing sampling steps and classifier-free guidance scales." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 313, + 72, + 553, + 227 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 72, + 553, + 227 + ], + "spans": [ + { + "bbox": [ + 313, + 72, + 553, + 227 + ], + "type": "text", + "content": "Metrics. We employ both objective and subjective evaluations across three dimensions: generation quality, audiovisual relevance, and audiovisual synchronization. For objective evaluation, generation quality is measured using FD score [21], KL divergence [15], and IS score [21]. In previous work, each metric may be computed with different feature extraction models (e.g., Melception, PasST" + }, + { + "bbox": [ + 313, + 72, + 553, + 227 + ], + "type": "inline_equation", + "content": "^1" + }, + { + "bbox": [ + 313, + 72, + 553, + 227 + ], + "type": "text", + "content": "). Audiovisual relevance is quantified by computing the cosine similarity between the embeddings of the input video and the generated audio using the ImageBind model [12]. For synchronization, we use the synchronization accuracy (Acc.) [25] reflects the alignment between the input video and generated audio." + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 313, + 228, + 553, + 443 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 228, + 553, + 443 + ], + "spans": [ + { + "bbox": [ + 313, + 228, + 553, + 443 + ], + "type": "text", + "content": "Baselines. Our method is comprehensively compared against all major generative paradigms, including autoregressive (AR), masked prediction (Mask), diffusion-based, and flow-based approaches, as detailed in Table 1. All baseline models were evaluated using either officially released code and models or directly downloaded audio outputs by their authors, with consistent experimental settings maintained across all evaluations on the same hardware platform. Implementation specifics include: (1) Im2Wav generates audio in 4s clips, truncated to 10s after three clips. (2) Diff-Foley here employs its double-guidance variant (demonstrating superior performance); (3) MMAudio-L here is the largest variant of MMAudio; and (4) Frieren here employs its final no-reflow configuration (achieving optimal Fréchet distance). For comprehensive evaluation, we also include text-augmented V2A models (denoted in gray rows) that leverage external textual data beyond benchmark audiovisual corpora during training or inference." + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 313, + 453, + 515, + 464 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 453, + 515, + 464 + ], + "spans": [ + { + "bbox": [ + 313, + 453, + 515, + 464 + ], + "type": "text", + "content": "4.2. Diffusion vs. Flow vs. VAFlow Models" + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 313, + 471, + 553, + 613 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 471, + 553, + 613 + ], + "spans": [ + { + "bbox": [ + 313, + 471, + 553, + 613 + ], + "type": "text", + "content": "We examine framework-specific performance by training same VAFlow-CLIP-142M models under three paradigms, maintaining identical configurations: CLIP visual features, architecture parameters, initialization weights, and other hyperparameters. The frameworks include: Diffusion: Standard DDPM [27, 33] with cross-attention video conditioning. Flow: Standard continuous-time flow (Equation 2) using Gaussian prior. VAFlow: Our framework employing aligned VAE-compressed video features as source distribution. All frameworks use their respective optimal inference settings (DDPM: DPMSolver; Flows: dopri5). Joint tuning in VAFlow begins at 100K steps." + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 313, + 615, + 553, + 674 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 615, + 553, + 674 + ], + "spans": [ + { + "bbox": [ + 313, + 615, + 553, + 674 + ], + "type": "text", + "content": "Frechet Distance (FD) results in Figure 3 reveals: 1) Flow paradigms converge faster with lower FD than diffusion, demonstrating inherent optimization benefits. 2) VAFlow attains lower final FD than standard flow despite delayed early-stage convergence (<100K steps). This per" + } + ] + } + ], + "index": 12 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 313, + 683, + 553, + 713 + ], + "type": "page_footnote", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 683, + 553, + 713 + ], + "spans": [ + { + "bbox": [ + 313, + 683, + 553, + 713 + ], + "type": "text", + "content": "1We adopt PasST [19] model over PANN [18] for its improved robustness following the practice introduced in AudioLDM [21] evaluation repository: https://github.com/haoheliu/audioldm_eval." + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "spans": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "type": "text", + "content": "11781" + } + ] + } + ], + "index": 14 + } + ], + "page_size": [ + 612, + 792 + ], + "page_idx": 4 + }, + { + "para_blocks": [ + { + "type": "image", + "bbox": [ + 85, + 72, + 267, + 190 + ], + "blocks": [ + { + "bbox": [ + 85, + 72, + 267, + 190 + ], + "lines": [ + { + "bbox": [ + 85, + 72, + 267, + 190 + ], + "spans": [ + { + "bbox": [ + 85, + 72, + 267, + 190 + ], + "type": "image", + "image_path": "ddbe06928bea183e718f5bdb4312d42104ea62742c41fc44089c0837a7187e51.jpg" + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 55, + 200, + 296, + 266 + ], + "lines": [ + { + "bbox": [ + 55, + 200, + 296, + 266 + ], + "spans": [ + { + "bbox": [ + 55, + 200, + 296, + 266 + ], + "type": "text", + "content": "Figure 3. Comparison of different training frameworks. We examine the same model under different training paradigms by controlling training data, steps, and other hyperparameters. Results demonstrate that VAFlow converges to superior optima, validating the effectiveness of its direct video-to-audio transformation design and highlighting this paradigm advantages in V2A generation." + } + ] + } + ], + "index": 1, + "angle": 0, + "type": "image_caption" + } + ], + "index": 0 + }, + { + "bbox": [ + 55, + 288, + 296, + 361 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 288, + 296, + 361 + ], + "spans": [ + { + "bbox": [ + 55, + 288, + 296, + 361 + ], + "type": "text", + "content": "formance attributes to VAFlow's explicit modeling of deterministic video-audio transport paths, avoiding the ambiguous noise-to-audio mappings in Gaussian-prior approaches. VAFlow learns optimal transport trajectories between video-audio distribution, bypassing error-prone intermediate transformations in Gaussian-noise-to-audio paths." + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 55, + 369, + 195, + 382 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 369, + 195, + 382 + ], + "spans": [ + { + "bbox": [ + 55, + 369, + 195, + 382 + ], + "type": "text", + "content": "4.3. VAFlow Scaling Analysis" + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 55, + 387, + 295, + 458 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 387, + 295, + 458 + ], + "spans": [ + { + "bbox": [ + 55, + 387, + 295, + 458 + ], + "type": "text", + "content": "We investigate VAFlow's scaling capabilities by developing four model variants (142M, 264M, 527M, 1.05B parameters) through width/depth scaling of DiT blocks while maintaining fixed training protocols (data, steps, etc.). Evaluation metrics (FD, IS) versus parameter counts are visualized in Figure 4 (bubble sizes denote model scales)." + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 55, + 459, + 295, + 639 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 459, + 295, + 639 + ], + "spans": [ + { + "bbox": [ + 55, + 459, + 295, + 639 + ], + "type": "text", + "content": "There are two key findings here: 1) Parameter Efficiency: VAFlow consistently outperforms flow-based baselines (Frieren, MMAudio) at comparable sizes (e.g., 142M vs. 159M, 1.05B vs. 1.03B) across both metrics. 2) Scaling Capability: Previous flow-based studies have struggled to show consistent scalability [8, 43], where larger models fail to achieve better metrics consistently. To the best of our knowledge, VAFlow is the first flow-based V2A framework to demonstrate stable scaling properties. As the parameter size increases, the models achieve better FD and IS scores. Interestingly, larger models tend to improve IS scores more significantly than FD scores, suggesting that as the VAFlow model scales, the quality of generated audio gradually saturates, and the model becomes more capable of deriving diverse audio from different video-sampled latents." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 55, + 647, + 170, + 659 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 647, + 170, + 659 + ], + "spans": [ + { + "bbox": [ + 55, + 647, + 170, + 659 + ], + "type": "text", + "content": "4.4. Benchmark Results" + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 55, + 665, + 296, + 715 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 665, + 296, + 715 + ], + "spans": [ + { + "bbox": [ + 55, + 665, + 296, + 715 + ], + "type": "text", + "content": "Quantitative evaluations are presented in Table 1, where VAFlow demonstrates state-of-the-art performance across key audio generation metrics: " + }, + { + "bbox": [ + 55, + 665, + 296, + 715 + ], + "type": "inline_equation", + "content": "\\mathrm{FD}_{\\text {pas. }}" + }, + { + "bbox": [ + 55, + 665, + 296, + 715 + ], + "type": "text", + "content": ", " + }, + { + "bbox": [ + 55, + 665, + 296, + 715 + ], + "type": "inline_equation", + "content": "\\mathrm{FD}_{\\text {mel. }}" + }, + { + "bbox": [ + 55, + 665, + 296, + 715 + ], + "type": "text", + "content": ", " + }, + { + "bbox": [ + 55, + 665, + 296, + 715 + ], + "type": "inline_equation", + "content": "\\mathrm{IS}_{\\text {pas. }}" + }, + { + "bbox": [ + 55, + 665, + 296, + 715 + ], + "type": "text", + "content": ", " + }, + { + "bbox": [ + 55, + 665, + 296, + 715 + ], + "type": "inline_equation", + "content": "\\mathrm{KL}_{\\text {pas. }}" + }, + { + "bbox": [ + 55, + 665, + 296, + 715 + ], + "type": "text", + "content": ", and " + }, + { + "bbox": [ + 55, + 665, + 296, + 715 + ], + "type": "inline_equation", + "content": "\\mathrm{KL}_{\\text {mel. }}" + }, + { + "bbox": [ + 55, + 665, + 296, + 715 + ], + "type": "text", + "content": ". For audio-visual synchronization," + } + ] + } + ], + "index": 7 + }, + { + "type": "image", + "bbox": [ + 344, + 72, + 526, + 207 + ], + "blocks": [ + { + "bbox": [ + 344, + 72, + 526, + 207 + ], + "lines": [ + { + "bbox": [ + 344, + 72, + 526, + 207 + ], + "spans": [ + { + "bbox": [ + 344, + 72, + 526, + 207 + ], + "type": "image", + "image_path": "f09fb5f06231f54fb4c79ca7f5346440e949c46f53aa2c0415e37e982158f883.jpg" + } + ] + } + ], + "index": 8, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 313, + 217, + 555, + 274 + ], + "lines": [ + { + "bbox": [ + 313, + 217, + 555, + 274 + ], + "spans": [ + { + "bbox": [ + 313, + 217, + 555, + 274 + ], + "type": "text", + "content": "Figure 4. Performance of flow-based baselines and VAFlow variants in scale. VAFlow models (142M, 1.05B) consistently outperform comparable-scale baselines (Frieren-159M, MMAudio-1.03B), with stable scaling property indicated by progressive IS gains (diversity) and saturating FD improvements (quality)." + } + ] + } + ], + "index": 9, + "angle": 0, + "type": "image_caption" + } + ], + "index": 8 + }, + { + "bbox": [ + 313, + 293, + 555, + 533 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 293, + 555, + 533 + ], + "spans": [ + { + "bbox": [ + 313, + 293, + 555, + 533 + ], + "type": "text", + "content": "VAFlow achieves " + }, + { + "bbox": [ + 313, + 293, + 555, + 533 + ], + "type": "inline_equation", + "content": "96.3\\%" + }, + { + "bbox": [ + 313, + 293, + 555, + 533 + ], + "type": "text", + "content": " accuracy (Acc.), comparable to Frieren " + }, + { + "bbox": [ + 313, + 293, + 555, + 533 + ], + "type": "inline_equation", + "content": "(97.0\\%)" + }, + { + "bbox": [ + 313, + 293, + 555, + 533 + ], + "type": "text", + "content": ". In semantic relevance, it attains an IB-VA score of 28.6, second only to TiVA's 30.3. Notably, VAFlow maintains a simple and straightforward model design without relying on complex video conditioning modules [8, 41, 46], carefully designed noise schedules [9], or additional guidance modules [25, 45]. Its performance across various metrics, achieved with a standard Flow Matching training approach and basic ODE solvers, underscores the simplicity and effectiveness of the direct video-to-audio mapping design. When compared to text-enhanced baselines (gray rows), VAFlow maintains advantages in " + }, + { + "bbox": [ + 313, + 293, + 555, + 533 + ], + "type": "inline_equation", + "content": "\\mathrm{FD}_{\\text{pas.}}" + }, + { + "bbox": [ + 313, + 293, + 555, + 533 + ], + "type": "text", + "content": ", " + }, + { + "bbox": [ + 313, + 293, + 555, + 533 + ], + "type": "inline_equation", + "content": "\\mathrm{FD}_{\\text{mel.}}" + }, + { + "bbox": [ + 313, + 293, + 555, + 533 + ], + "type": "text", + "content": ", " + }, + { + "bbox": [ + 313, + 293, + 555, + 533 + ], + "type": "inline_equation", + "content": "\\mathrm{IS}_{\\text{pas.}}" + }, + { + "bbox": [ + 313, + 293, + 555, + 533 + ], + "type": "text", + "content": ", and synchronization Acc., though text-augmented models exhibit superior audio semantic metrics (KL distances and IB-VA). This aligns with expectations since text conditioning explicitly constrains semantic space. We note that Seeing&Hearing's elevated IB-VA score (36.7) stems from classifier guidance using the same ImageBind model, which however compromises generation quality and synchronization performance." + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 313, + 533, + 556, + 653 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 533, + 556, + 653 + ], + "spans": [ + { + "bbox": [ + 313, + 533, + 556, + 653 + ], + "type": "text", + "content": "We further conduct human evaluation with sampled 50 videos from the test set to compare our method with the flow-based baseline, Frieren. The generated results from both models were randomized, and 10 experts were invited to score the models based on three criteria: sound quality (Quality), audio-visual semantic consistency (Semantic) and audio-visual synchronization (Sync.). The scoring was done by selecting a winner or declaring a tie for each comparison. The results, summarized in Table 2, show that our model outperforms the baseline in all three dimensions." + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 313, + 659, + 419, + 673 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 659, + 419, + 673 + ], + "spans": [ + { + "bbox": [ + 313, + 659, + 419, + 673 + ], + "type": "text", + "content": "4.5. Ablative Analysis" + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 313, + 677, + 555, + 715 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 677, + 555, + 715 + ], + "spans": [ + { + "bbox": [ + 313, + 677, + 555, + 715 + ], + "type": "text", + "content": "Visual Features. Three visual feature types were experimented with: semantic features (CLIP embeddings extracted per frame at 10 fps), temporally-enhanced semantic" + } + ] + } + ], + "index": 13 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 294, + 748, + 318, + 757 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 294, + 748, + 318, + 757 + ], + "spans": [ + { + "bbox": [ + 294, + 748, + 318, + 757 + ], + "type": "text", + "content": "11782" + } + ] + } + ], + "index": 14 + } + ], + "page_size": [ + 612, + 792 + ], + "page_idx": 5 + }, + { + "para_blocks": [ + { + "type": "table", + "bbox": [ + 58, + 70, + 556, + 301 + ], + "blocks": [ + { + "bbox": [ + 58, + 70, + 556, + 301 + ], + "lines": [ + { + "bbox": [ + 58, + 70, + 556, + 301 + ], + "spans": [ + { + "bbox": [ + 58, + 70, + 556, + 301 + ], + "type": "table", + "html": "
ParadigmMethodVis.Feat.QualitySync. Acc. ↑Semantic IB-VA ↑
\\(FD_{pas.}\\)↓\\(FD_{mel.}\\)↓\\(IS_{mel.}\\)↑\\(KL_{pas.}\\)↓\\(KL_{mel.}\\)↓
ARSpecVQGAN [15]RGB,Opti.342.418.1520.53.106.8455.914.1
Im2Wav [34]CLIP252.213.0335.82.265.4178.919.6
MaskVATT [24]eva-CLIP+154.17.0764.31.343.4684.825.0
VAB [37]eva-CLIP193.112.2134.42.315.0982.425.9
DiffusionMultiFoley [7]CAVP241.613.0067.71.593.3485.227.0
Seeing&Hearing [45]ImageBind+251.512.630.12.656.3754.236.7
Diff-Foley [25]CAVP512.111.4550.42.916.1888.020.7
LoVA [9]CLIP149.16.9058.72.105.1287.026.3
FoleyCrafter [46]CLIP134.27.4056.62.295.6583.527.8
TiVA [41]CLIP111.94.3167.02.025.2187.530.3
V2A-Mapper [40]CLIP94.94.2164.92.466.0678.622.2
FlowMMAudio-L [8]CLIP+159.29.3845.41.953.8477.533.9
Frieren [43]CAVP259.96.7052.42.925.6497.023.0
VAFlow (ours)CLIP87.73.8673.61.914.8188.128.6
VAFlow (ours)CAVP91.84.7364.82.415.9696.325.1
VAFlow (ours)VidTok128.56.1748.13.648.9449.912.9
", + "image_path": "b5c13a95ac5535f0a7bfdf235fa260cbc27c52e8be3881895d050442f03d78b9.jpg" + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "table_body" + } + ], + "index": 0 + }, + { + "type": "table", + "bbox": [ + 59, + 409, + 292, + 453 + ], + "blocks": [ + { + "bbox": [ + 55, + 309, + 555, + 390 + ], + "lines": [ + { + "bbox": [ + 55, + 309, + 555, + 390 + ], + "spans": [ + { + "bbox": [ + 55, + 309, + 555, + 390 + ], + "type": "text", + "content": "Table 1. Quantitative video-to-audio results on the VGGSound test set. “Vis. Feat.” indicates the employed visual features, with a “+” denoting the inclusion of auxiliary visual information (e.g., visual captions via LLama-2 [39] or Qwen-VL [2] in VATT and Seeing&Hearing, or SyncFormer [16] features in MMAudio). All results are obtained from official code or released audio outputs. The MultiFoley results here is an 8k subset filtered by ImageBind from the authors. Rows in gray denote methods that incorporate extra text data (e.g., video captions); for fairness, these results are provided for reference only and are not directly compared with standard video-to-audio models. The best score is bolded, and the second-best is underlined. Our three VAFlow variants, differing in visual features, achieve definitive gains in audio quality (FD, IS, KL) while delivering near state-of-the-art performance for synchronization (Acc.) and correlation (IB-VA)." + } + ] + } + ], + "index": 1, + "angle": 0, + "type": "table_caption" + }, + { + "bbox": [ + 59, + 409, + 292, + 453 + ], + "lines": [ + { + "bbox": [ + 59, + 409, + 292, + 453 + ], + "spans": [ + { + "bbox": [ + 59, + 409, + 292, + 453 + ], + "type": "table", + "html": "
ComparisonQualityWin rate(%) Sync.Semantic
Ours vs. Frieren66.00(±4.18)60.44(±5.72)60.89(±4.48)
", + "image_path": "cb0cc149a1b8270bb5b2fdc91aff8972184a97bd26578f2298e92afecff4b222.jpg" + } + ] + } + ], + "index": 2, + "angle": 0, + "type": "table_body" + } + ], + "index": 2 + }, + { + "bbox": [ + 55, + 461, + 295, + 506 + ], + "lines": [ + { + "bbox": [ + 55, + 461, + 295, + 506 + ], + "spans": [ + { + "bbox": [ + 55, + 461, + 295, + 506 + ], + "type": "text", + "content": "Table 2. Human evaluation results, with comparison between our model and the flow-based baseline Frieren across three criteria: sound quality, audio-visual synchronization, and audio-visual semantic consistency. Ours outperforms the baseline in all criteria." + } + ] + } + ], + "index": 3, + "angle": 0, + "type": "text" + }, + { + "bbox": [ + 54, + 533, + 295, + 713 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 54, + 533, + 295, + 713 + ], + "spans": [ + { + "bbox": [ + 54, + 533, + 295, + 713 + ], + "type": "text", + "content": "features (CAVP features via video-audio contrastive learning at 4 fps), reconstruction features (VidTok tokenized latents at 15 fps). Utilizing the same model size and training process for these features, we developed three variants: VAFlow-CLIP, VAFlow-CAVP, and VAFlow-VidTok, with results presented in the last three rows of Table 1. The results indicate that the widely-used CLIP features overall perform the best, yielding superior audio quality and semantic relevance. The temporally enhanced CAVP features, on the other hand, demonstrate the best video-audio synchronization but slightly lag in other metrics compared to VAFlow-CLIP. Models based on reconstruction features exhibit the poorest performance across all evaluated aspects. These outcomes suggest two key insights: 1) The type of video features directly influences VAFlow's performance," + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 313, + 411, + 555, + 507 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 411, + 555, + 507 + ], + "spans": [ + { + "bbox": [ + 313, + 411, + 555, + 507 + ], + "type": "text", + "content": "with different feature types leading to models with distinct performance preferences (either better semantic performance or synchronization); 2) Modality alignment matters more than information preservation. Although reconstruction features retain the most comprehensive video information among the three, their lack of modality alignment hampers the V2A model's ability to learn effective cross-modal generative capabilities." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 313, + 508, + 556, + 592 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 508, + 556, + 592 + ], + "spans": [ + { + "bbox": [ + 313, + 508, + 556, + 592 + ], + "type": "text", + "content": "Flow Matching Source Distributions. We evaluate VAFlow under two source distributions: video latent priors from alignment VAE and standard Gaussian noise. For video prior, we further test variants with/without joint training of the alignment VAE encoder (Method 3.3). All models share same training budgets (300K steps for non-joint training variants; 170K+130K for joint training)." + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 313, + 593, + 556, + 689 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 593, + 556, + 689 + ], + "spans": [ + { + "bbox": [ + 313, + 593, + 556, + 689 + ], + "type": "text", + "content": "Results in Table 3 demonstrate that video latent priors trained solely under alignment VAE objective in Equation 7 fail to surpass Gaussian noise baselines. Only when coupled with joint training—adapting the prior to flow-matching tasks—do video latent priors outperform Gaussian prior. This highlights the necessity of co-optimizing prior distributions for flow-based generation rather than relying on standalone pretraining on reconstruction task." + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 313, + 689, + 555, + 715 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 689, + 555, + 715 + ], + "spans": [ + { + "bbox": [ + 313, + 689, + 555, + 715 + ], + "type": "text", + "content": "Conditioning Mechanisms and Guidance. Prior work highlights the critical role of condition and guidance mech" + } + ] + } + ], + "index": 8 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 294, + 748, + 318, + 757 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 294, + 748, + 318, + 757 + ], + "spans": [ + { + "bbox": [ + 294, + 748, + 318, + 757 + ], + "type": "text", + "content": "11783" + } + ] + } + ], + "index": 9 + } + ], + "page_size": [ + 612, + 792 + ], + "page_idx": 6 + }, + { + "para_blocks": [ + { + "type": "table", + "bbox": [ + 66, + 70, + 287, + 125 + ], + "blocks": [ + { + "bbox": [ + 66, + 70, + 287, + 125 + ], + "lines": [ + { + "bbox": [ + 66, + 70, + 287, + 125 + ], + "spans": [ + { + "bbox": [ + 66, + 70, + 287, + 125 + ], + "type": "table", + "html": "
V-PriorTuned\\(FD_{pas.}\\)↓\\(IS_{mel.}\\)↑Acc.↑IB-VA ↑
X-102.667.687.927.9
X111.857.482.325.4
87.773.688.128.6
", + "image_path": "72a285c0140b286cc05b89f5a361a88df831d6a8cdedaf0be2c027104301c080.jpg" + } + ] + } + ], + "index": 0, + "angle": 0, + "type": "table_body" + } + ], + "index": 0 + }, + { + "type": "table", + "bbox": [ + 59, + 213, + 293, + 277 + ], + "blocks": [ + { + "bbox": [ + 55, + 133, + 295, + 189 + ], + "lines": [ + { + "bbox": [ + 55, + 133, + 295, + 189 + ], + "spans": [ + { + "bbox": [ + 55, + 133, + 295, + 189 + ], + "type": "text", + "content": "Table 3. Ablation study on flow-matching source distributions. 'V-Prior' denotes sampling starts from video latent priors (vs. Gaussian priors). 'Tuned' indicates joint optimization of the alignment VAE encoder. Results show that jointly tuned video priors outperform Gaussian priors as source distributions for V2A flow models." + } + ] + } + ], + "index": 1, + "angle": 0, + "type": "table_caption" + }, + { + "bbox": [ + 59, + 213, + 293, + 277 + ], + "lines": [ + { + "bbox": [ + 59, + 213, + 293, + 277 + ], + "spans": [ + { + "bbox": [ + 59, + 213, + 293, + 277 + ], + "type": "table", + "html": "
Cond.CACFGFDpas. ↓ISmel. ↑Acc. ↑IB-VA ↑
XX-135.941.431.78.5
XX160.536.518.95.4
X133.842.378.021.0
87.773.688.128.6
", + "image_path": "b3747903e3af065be5805bc844baf66fb78cff9f80dc36e605eea64dee29096e.jpg" + } + ] + } + ], + "index": 2, + "angle": 0, + "type": "table_body" + } + ], + "index": 2 + }, + { + "bbox": [ + 55, + 286, + 295, + 353 + ], + "lines": [ + { + "bbox": [ + 55, + 286, + 295, + 353 + ], + "spans": [ + { + "bbox": [ + 55, + 286, + 295, + 353 + ], + "type": "text", + "content": "Table 4. Ablation study on conditioning mechanisms and guidance. 'Cond.' denotes explicit video conditioning during inference; 'CA' indicates the presence of cross-attention layers of different variants; 'CFG' refers to classifier-free guidance. Results demonstrate that optimal performance requires both explicit conditioning and CFG when solving VAFlow ODE." + } + ] + } + ], + "index": 3, + "angle": 0, + "type": "text" + }, + { + "bbox": [ + 55, + 388, + 295, + 413 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 388, + 295, + 413 + ], + "spans": [ + { + "bbox": [ + 55, + 388, + 295, + 413 + ], + "type": "text", + "content": "anisms (e.g., classifier-free guidance, CFG) in enhancing model performance. We address two questions:" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 55, + 419, + 296, + 714 + ], + "type": "list", + "angle": 0, + "index": 7, + "blocks": [ + { + "bbox": [ + 55, + 419, + 296, + 622 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 419, + 296, + 622 + ], + "spans": [ + { + "bbox": [ + 55, + 419, + 296, + 622 + ], + "type": "text", + "content": "(1) Can VAFlow generate high-quality audio using only video priors, omitting explicit video conditioning during sampling? We experimented with two variants without explicit condition: No cross-attention (CA): Remove CA layers in DiT blocks and retrain the model (Table 4, Row 1). Zero-input CA: Retain CA layers but zero out video inputs during inference (Table 4, Row 2). 'No CA' variant exhibit comparable audio quality metrics (FD, IS) to the video-condition-visible variant (Table 4, Row 3: explicit video conditioning without CFG), yet suffer significant degradation in semantic relevance (IB-VA) and synchronization (Acc.). This indicates that while video priors partially preserve visual information, explicit video conditioning during sampling is essential for temporal and semantic alignment. We think that as the latent trajectory approaches the target audio distribution " + }, + { + "bbox": [ + 55, + 419, + 296, + 622 + ], + "type": "inline_equation", + "content": "(t\\to 1)" + }, + { + "bbox": [ + 55, + 419, + 296, + 622 + ], + "type": "text", + "content": ", initial visual cues in the prior degrade, necessitating direct video information." + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 55, + 629, + 296, + 714 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 629, + 296, + 714 + ], + "spans": [ + { + "bbox": [ + 55, + 629, + 296, + 714 + ], + "type": "text", + "content": "(2) Does explicit video conditioning during sampling still require CFG for optimal performance? Adding CFG to the video-condition-visible variant (Row 4 in Table 4) further improves all metrics, demonstrating VAFlow's compatibility with established guidance techniques. This underscores the dual necessity of explicit video conditioning and CFG for optimal V2A synthesis with VAFlow framework." + } + ] + } + ], + "index": 6 + } + ], + "sub_type": "text" + }, + { + "type": "image", + "bbox": [ + 316, + 94, + 553, + 356 + ], + "blocks": [ + { + "bbox": [ + 323, + 70, + 364, + 78 + ], + "lines": [ + { + "bbox": [ + 323, + 70, + 364, + 78 + ], + "spans": [ + { + "bbox": [ + 323, + 70, + 364, + 78 + ], + "type": "text", + "content": "A case video:" + } + ] + } + ], + "index": 8, + "angle": 0, + "type": "image_caption" + }, + { + "bbox": [ + 324, + 78, + 543, + 93 + ], + "lines": [ + { + "bbox": [ + 324, + 78, + 543, + 93 + ], + "spans": [ + { + "bbox": [ + 324, + 78, + 543, + 93 + ], + "type": "text", + "content": "At a seaside with rolling waves, a group of people are swimming and playing. A woman walks toward the camera, points to the right, and says something." + } + ] + } + ], + "index": 9, + "angle": 0, + "type": "image_caption" + }, + { + "bbox": [ + 316, + 94, + 553, + 356 + ], + "lines": [ + { + "bbox": [ + 316, + 94, + 553, + 356 + ], + "spans": [ + { + "bbox": [ + 316, + 94, + 553, + 356 + ], + "type": "image", + "image_path": "541834df9a135b3f673f2135fc672b6335d888f01e8ee58e170b788fb154322c.jpg" + } + ] + } + ], + "index": 10, + "angle": 0, + "type": "image_body" + }, + { + "bbox": [ + 313, + 363, + 555, + 407 + ], + "lines": [ + { + "bbox": [ + 313, + 363, + 555, + 407 + ], + "spans": [ + { + "bbox": [ + 313, + 363, + 555, + 407 + ], + "type": "text", + "content": "Figure 5. Mel-spectrograms of audio generated by VAFlow and the baseline for a complex seaside scene, including background waves and varying voices, demonstrating VAFlow's superior content understanding and synchronization with the video." + } + ] + } + ], + "index": 11, + "angle": 0, + "type": "image_caption" + } + ], + "index": 10 + }, + { + "bbox": [ + 314, + 431, + 397, + 441 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 314, + 431, + 397, + 441 + ], + "spans": [ + { + "bbox": [ + 314, + 431, + 397, + 441 + ], + "type": "text", + "content": "4.6. Case Studies" + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 313, + 449, + 555, + 570 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 449, + 555, + 570 + ], + "spans": [ + { + "bbox": [ + 313, + 449, + 555, + 570 + ], + "type": "text", + "content": "Figure 5 illustrates the performance of VAFlow and the baselines on the complex scene video. The video depicts a seaside scene with various sounds include background waves and different voices associated with shifting characters (noisy laughter, woman speaking). We present the ground truth and the mel-spectrograms of the audio generated by different methods. Our results show that VAFlow not only accurately understands the video content and generates all the necessary sounds but also maintains synchronization with the visual timing." + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 313, + 584, + 479, + 596 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 584, + 479, + 596 + ], + "spans": [ + { + "bbox": [ + 313, + 584, + 479, + 596 + ], + "type": "text", + "content": "5. Conclusion and Future Work" + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 313, + 605, + 555, + 715 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 313, + 605, + 555, + 715 + ], + "spans": [ + { + "bbox": [ + 313, + 605, + 555, + 715 + ], + "type": "text", + "content": "In this paper, we introduce VAFlow, a novel framework for V2A task, which directly denoises from the video space to the audio space via flow matching. demonstrating superior performance compared to previous methods. Future work will explore VAFlow's potential in more diverse audio domains, such as speech and music. Expanding the range of tasks and datasets will enable VAFlow's scaling properties to shine even further, laying a more general-purpose the foundation model for content generation community." + } + ] + } + ], + "index": 15 + } + ], + "discarded_blocks": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "spans": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "type": "text", + "content": "11784" + } + ] + } + ], + "index": 16 + } + ], + "page_size": [ + 612, + 792 + ], + "page_idx": 7 + }, + { + "para_blocks": [ + { + "bbox": [ + 56, + 72, + 153, + 85 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 72, + 153, + 85 + ], + "spans": [ + { + "bbox": [ + 56, + 72, + 153, + 85 + ], + "type": "text", + "content": "Acknowledgement" + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 55, + 91, + 296, + 139 + ], + "type": "text", + "angle": 0, + "lines": [ + { + "bbox": [ + 55, + 91, + 296, + 139 + ], + "spans": [ + { + "bbox": [ + 55, + 91, + 296, + 139 + ], + "type": "text", + "content": "This work is supported by the National Natural Science Foundation of China (No. 62276268), Public Computing Cloud, Renmin University of China and ZHI-TECH GROUP." + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 56, + 148, + 115, + 161 + ], + "type": "title", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 148, + 115, + 161 + ], + "spans": [ + { + "bbox": [ + 56, + 148, + 115, + 161 + ], + "type": "text", + "content": "References" + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 57, + 169, + 296, + 713 + ], + "type": "list", + "angle": 0, + "index": 15, + "blocks": [ + { + "bbox": [ + 61, + 169, + 296, + 201 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 61, + 169, + 296, + 201 + ], + "spans": [ + { + "bbox": [ + 61, + 169, + 296, + 201 + ], + "type": "text", + "content": "[1] Michael S. Albergo and Eric Vanden-Eijnden. Building normalizing flows with stochastic interpolants. In ICLR, 2023. 2" + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 61, + 202, + 296, + 258 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 61, + 202, + 296, + 258 + ], + "spans": [ + { + "bbox": [ + 61, + 202, + 296, + 258 + ], + "type": "text", + "content": "[2] Jinze Bai, Shuai Bai, Shusheng Yang, Shijie Wang, Sinan Tan, Peng Wang, Junyang Lin, Chang Zhou, and Jingren Zhou. Qwen-vl: A versatile vision-language model for understanding, localization, text reading, and beyond. arXiv preprint arXiv:2308.12966, 2023. 7" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 61, + 258, + 295, + 312 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 61, + 258, + 295, + 312 + ], + "spans": [ + { + "bbox": [ + 61, + 258, + 295, + 312 + ], + "type": "text", + "content": "[3] Tim Brooks, Bill Peebles, Connor Holmes, Will DePue, Yufei Guo, Li Jing, David Schnurr, Joe Taylor, Troy Luhman, Eric Luhman, Clarence Ng, Ricky Wang, and Aditya Ramesh. Video generation models as world simulators, 2024. 1" + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 61, + 313, + 295, + 369 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 61, + 313, + 295, + 369 + ], + "spans": [ + { + "bbox": [ + 61, + 313, + 295, + 369 + ], + "type": "text", + "content": "[4] Li-Wen Chang, Wenlei Bao, Qi Hou, Chengquan Jiang, Ningxin Zheng, Yinmin Zhong, Xuanrun Zhang, Zuquan Song, Chengji Yao, Ziheng Jiang, et al. Flux: fast software-based communication overlap on gpus through kernel fusion. arXiv preprint arXiv:2406.06858, 2024. 1" + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 62, + 369, + 295, + 402 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 62, + 369, + 295, + 402 + ], + "spans": [ + { + "bbox": [ + 62, + 369, + 295, + 402 + ], + "type": "text", + "content": "[5] Honglie Chen, Weidi Xie, Andrea Vedaldi, and Andrew Zisserman. Vggsound: A large-scale audio-visual dataset. In ICASSP, pages 721–725, 2020. 5" + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 62, + 403, + 295, + 446 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 62, + 403, + 295, + 446 + ], + "spans": [ + { + "bbox": [ + 62, + 403, + 295, + 446 + ], + "type": "text", + "content": "[6] Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A simple framework for contrastive learning of visual representations. In International Conference on Machine Learning, pages 1597-1607, 2020. 4" + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 62, + 447, + 295, + 491 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 62, + 447, + 295, + 491 + ], + "spans": [ + { + "bbox": [ + 62, + 447, + 295, + 491 + ], + "type": "text", + "content": "[7] Ziyang Chen, Prem Seetharaman, Bryan Russell, Oriol Nieto, David Bourgin, Andrew Owens, and Justin Salamon. Video-guided foley sound generation with multimodal controls. arXiv preprint arXiv:2411.17698, 2024. 2, 7" + } + ] + } + ], + "index": 9 + }, + { + "bbox": [ + 62, + 491, + 295, + 536 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 62, + 491, + 295, + 536 + ], + "spans": [ + { + "bbox": [ + 62, + 491, + 295, + 536 + ], + "type": "text", + "content": "[8] Ho Kei Cheng, Masato Ishii, Akio Hayakawa, Takashi Shibuya, Alexander Schwing, and Yuki Mitsufuji. Taming multimodal joint training for high-quality video-to-audio synthesis. arXiv preprint arXiv:2412.15322, 2024. 2, 3, 6, 7" + } + ] + } + ], + "index": 10 + }, + { + "bbox": [ + 62, + 536, + 295, + 568 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 62, + 536, + 295, + 568 + ], + "spans": [ + { + "bbox": [ + 62, + 536, + 295, + 568 + ], + "type": "text", + "content": "[9] Xin Cheng, Xihua Wang, Yihan Wu, Yuyue Wang, and Ruihua Song. Lova: Long-form video-to-audio generation. arXiv preprint arXiv:2409.15157, 2024. 1, 2, 6, 7" + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 57, + 570, + 295, + 634 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 57, + 570, + 295, + 634 + ], + "spans": [ + { + "bbox": [ + 57, + 570, + 295, + 634 + ], + "type": "text", + "content": "[10] Patrick Esser, Sumith Kulal, Andreas Blattmann, Rahim Entezari, Jonas Müller, Harry Saini, Yam Levi, Dominik Lorenz, Axel Sauer, Frederic Boesel, et al. Scaling rectified flow transformers for high-resolution image synthesis. In International Conference on Machine Learning, pages 12606-12633, 2024. 1" + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 57, + 635, + 295, + 668 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 57, + 635, + 295, + 668 + ], + "spans": [ + { + "bbox": [ + 57, + 635, + 295, + 668 + ], + "type": "text", + "content": "[11] Zach Evans, Julian D Parker, CJ Carr, Zack Zukowski, Josiah Taylor, and Jordi Pons. Stable audio open. arXiv preprint arXiv:2407.14358, 2024. 3, 5" + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 57, + 670, + 295, + 713 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 57, + 670, + 295, + 713 + ], + "spans": [ + { + "bbox": [ + 57, + 670, + 295, + 713 + ], + "type": "text", + "content": "[12] Rohit Girdhar, Alaaeldin El-Nouby, Zhuang Liu, Mannat Singh, Kalyan Vasudev Alwala, Armand Joulin, and Ishan Misra. Imagebind: One embedding space to bind them all. In CVPR, pages 15180-15190, 2023. 5" + } + ] + } + ], + "index": 14 + } + ], + "sub_type": "ref_text" + }, + { + "bbox": [ + 316, + 73, + 554, + 712 + ], + "type": "list", + "angle": 0, + "index": 31, + "blocks": [ + { + "bbox": [ + 316, + 73, + 554, + 127 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 73, + 554, + 127 + ], + "spans": [ + { + "bbox": [ + 316, + 73, + 554, + 127 + ], + "type": "text", + "content": "[13] Aaron Grattafori, Abhimanyu Dubey, Abhinav Jauhri, Abhinav Pandey, Abhishek Kadian, Ahmad Al-Dahle, Aiesha Letman, Akhil Mathur, Alan Schelten, Alex Vaughan, et al. The llama 3 herd of models. arXiv preprint arXiv:2407.21783, 2024. 2" + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 316, + 129, + 553, + 150 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 129, + 553, + 150 + ], + "spans": [ + { + "bbox": [ + 316, + 129, + 553, + 150 + ], + "type": "text", + "content": "[14] Jonathan Ho and Tim Salimans. Classifier-free diffusion guidance. arXiv preprint arXiv:2207.12598, 2022. 5" + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 316, + 152, + 553, + 175 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 152, + 553, + 175 + ], + "spans": [ + { + "bbox": [ + 316, + 152, + 553, + 175 + ], + "type": "text", + "content": "[15] Vladimir iashin and Esa Rahtu. Taming visually guided sound generation. In BMVC, 2021. 1, 2, 5, 7" + } + ] + } + ], + "index": 18 + }, + { + "bbox": [ + 316, + 175, + 553, + 208 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 175, + 553, + 208 + ], + "spans": [ + { + "bbox": [ + 316, + 175, + 553, + 208 + ], + "type": "text", + "content": "[16] Vladimir Iashin, Weidi Xie, Esa Rahtu, and Andrew Zisserman. Synchformer: Efficient synchronization from sparse cues. In ICASSP, pages 5325-5329, 2024. 7" + } + ] + } + ], + "index": 19 + }, + { + "bbox": [ + 316, + 209, + 553, + 253 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 209, + 553, + 253 + ], + "spans": [ + { + "bbox": [ + 316, + 209, + 553, + 253 + ], + "type": "text", + "content": "[17] Tero Karras, Miika Aittala, Timo Aila, and Samuli Laine. Elucidating the design space of diffusion-based generative models. In Advances in Neural Information Processing Systems, pages 26565-26577, 2022. 2" + } + ] + } + ], + "index": 20 + }, + { + "bbox": [ + 316, + 254, + 553, + 308 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 254, + 553, + 308 + ], + "spans": [ + { + "bbox": [ + 316, + 254, + 553, + 308 + ], + "type": "text", + "content": "[18] Qiuqiang Kong, Yin Cao, Turab Iqbal, Yuxuan Wang, Wenwu Wang, and Mark D Plumbley. Panns: Large-scale pretrained audio neural networks for audio pattern recognition. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 28:2880-2894, 2020. 5" + } + ] + } + ], + "index": 21 + }, + { + "bbox": [ + 316, + 310, + 553, + 343 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 310, + 553, + 343 + ], + "spans": [ + { + "bbox": [ + 316, + 310, + 553, + 343 + ], + "type": "text", + "content": "[19] Khaled Koutini, Jan Schlüter, Hamid Eghbal-Zadeh, and Gerhard Widmer. Efficient training of audio transformers with patchout. arXiv preprint arXiv:2110.05069, 2021. 5" + } + ] + } + ], + "index": 22 + }, + { + "bbox": [ + 316, + 344, + 553, + 376 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 344, + 553, + 376 + ], + "spans": [ + { + "bbox": [ + 316, + 344, + 553, + 376 + ], + "type": "text", + "content": "[20] Yaron Lipman, Ricky T. Q. Chen, Heli Ben-Hamu, Maximilian Nickel, and Matthew Le. Flow matching for generative modeling. In ICLR, 2023. 2, 3" + } + ] + } + ], + "index": 23 + }, + { + "bbox": [ + 316, + 378, + 553, + 432 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 378, + 553, + 432 + ], + "spans": [ + { + "bbox": [ + 316, + 378, + 553, + 432 + ], + "type": "text", + "content": "[21] Haohe Liu, Zehua Chen, Yi Yuan, Xinhao Mei, Xubo Liu, Danilo Mandic, Wenwu Wang, and Mark D Plumbley. Audioldm: Text-to-audio generation with latent diffusion models. In International Conference on Machine Learning, pages 21450-21474, 2023. 5" + } + ] + } + ], + "index": 24 + }, + { + "bbox": [ + 316, + 434, + 553, + 477 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 434, + 553, + 477 + ], + "spans": [ + { + "bbox": [ + 316, + 434, + 553, + 477 + ], + "type": "text", + "content": "[22] Qihao Liu, Xi Yin, Alan Yuille, Andrew Brown, and Mannat Singh. Flowing from words to pixels: A noise-free framework for cross-modality evolution. In CVPR, pages 2755-2765, 2025. 2" + } + ] + } + ], + "index": 25 + }, + { + "bbox": [ + 316, + 479, + 553, + 511 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 479, + 553, + 511 + ], + "spans": [ + { + "bbox": [ + 316, + 479, + 553, + 511 + ], + "type": "text", + "content": "[23] Xingchao Liu, Chengyue Gong, and Qiang Liu. Flow straight and fast: Learning to generate and transfer data with rectified flow. In ICLR, 2023. 2, 3" + } + ] + } + ], + "index": 26 + }, + { + "bbox": [ + 316, + 512, + 553, + 556 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 512, + 553, + 556 + ], + "spans": [ + { + "bbox": [ + 316, + 512, + 553, + 556 + ], + "type": "text", + "content": "[24] Xiulong Liu, Kun Su, and Eli Shlizerman. Tell what you hear from what you see - video to audio generation through text. In Advances in Neural Information Processing Systems, pages 101337-101366, 2024. 1, 2, 7" + } + ] + } + ], + "index": 27 + }, + { + "bbox": [ + 316, + 557, + 553, + 612 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 557, + 553, + 612 + ], + "spans": [ + { + "bbox": [ + 316, + 557, + 553, + 612 + ], + "type": "text", + "content": "[25] Simian Luo, Chuanhao Yan, Chenxu Hu, and Hang Zhao. Diff-foley: Synchronized video-to-audio synthesis with latent diffusion models. In Advances in Neural Information Processing Systems, pages 48855-48876, 2023. 1, 2, 3, 5, 6, 7" + } + ] + } + ], + "index": 28 + }, + { + "bbox": [ + 316, + 613, + 553, + 668 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 613, + 553, + 668 + ], + "spans": [ + { + "bbox": [ + 316, + 613, + 553, + 668 + ], + "type": "text", + "content": "[26] Xinhao Mei, Varun Nagaraja, Gael Le Lan, Zhaoheng Ni, Ernie Chang, Yangyang Shi, and Vikas Chandra. Foleygen: Visually-guided audio generation. In 2024 IEEE 34th International Workshop on Machine Learning for Signal Processing, pages 1-6, 2024. 1, 2, 3" + } + ] + } + ], + "index": 29 + }, + { + "bbox": [ + 316, + 669, + 553, + 712 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 669, + 553, + 712 + ], + "spans": [ + { + "bbox": [ + 316, + 669, + 553, + 712 + ], + "type": "text", + "content": "[27] Alexander Quinn Nichol and Prafulla Dhariwal. Improved denoising diffusion probabilistic models. In International Conference on Machine Learning, pages 8162-8171, 2021. 2, 5" + } + ] + } + ], + "index": 30 + } + ], + "sub_type": "ref_text" + } + ], + "discarded_blocks": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "spans": [ + { + "bbox": [ + 294, + 748, + 317, + 757 + ], + "type": "text", + "content": "11785" + } + ] + } + ], + "index": 32 + } + ], + "page_size": [ + 612, + 792 + ], + "page_idx": 8 + }, + { + "para_blocks": [ + { + "bbox": [ + 56, + 72, + 294, + 713 + ], + "type": "list", + "angle": 0, + "index": 10, + "blocks": [ + { + "bbox": [ + 56, + 72, + 294, + 116 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 72, + 294, + 116 + ], + "spans": [ + { + "bbox": [ + 56, + 72, + 294, + 116 + ], + "type": "text", + "content": "[28] Shen Nie, Fengqi Zhu, Chao Du, Tianyu Pang, Qian Liu, Guangtao Zeng, Min Lin, and Chongxuan Li. Scaling up masked diffusion models on text. arXiv preprint arXiv:2410.18514, 2024. 2" + } + ] + } + ], + "index": 0 + }, + { + "bbox": [ + 56, + 118, + 294, + 162 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 118, + 294, + 162 + ], + "spans": [ + { + "bbox": [ + 56, + 118, + 294, + 162 + ], + "type": "text", + "content": "[29] Santiago Pascual, Chunghsin Yeh, Ioannis Tsiamas, and Joan Serra. Masked generative video-to-audio transformers with enhanced synchronicity. In ECCV, pages 247-264, 2024. 1, 2, 3" + } + ] + } + ], + "index": 1 + }, + { + "bbox": [ + 57, + 164, + 294, + 186 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 57, + 164, + 294, + 186 + ], + "spans": [ + { + "bbox": [ + 57, + 164, + 294, + 186 + ], + "type": "text", + "content": "[30] William Peebles and Saining Xie. Scalable diffusion models with transformers. In ICCV, pages 4195-4205, 2023. 2" + } + ] + } + ], + "index": 2 + }, + { + "bbox": [ + 56, + 188, + 294, + 449 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 188, + 294, + 449 + ], + "spans": [ + { + "bbox": [ + 56, + 188, + 294, + 449 + ], + "type": "text", + "content": "[31] Adam Polyak, Amit Zohar, Andrew Brown, Andros Tjandra, Animesh Sinha, Ann Lee, Apoorv Vyas, Bowen Shi, Chih-Yao Ma, Ching-Yao Chuang, David Yan, Dhruv Choudhary, Dingkang Wang, Geet Sethi, Guan Pang, Haoyu Ma, Ishan Misra, Ji Hou, Jialiang Wang, Kiran Jagadeesh, Kunpeng Li, Luxin Zhang, Mannat Singh, Mary Williamson, Matt Le, Matthew Yu, Mitesh Kumar Singh, Peizhao Zhang, Peter Vajda, Quentin Duval, Rohit Girdhar, Roshan Sumbaly, Sai Saketh Rambhatla, Sam Tsai, Samaneh Azadi, Samyak Datta, Sanyuan Chen, Sean Bell, Sharadh Ramaswamy, Shelly Sheynin, Siddharth Bhattacharya, Simran Motwani, Tao Xu, Tianhe Li, Tingbo Hou, Wei-Ning Hsu, Xi Yin, Xiaoliang Dai, Yaniv Taigman, Yaqiao Luo, Yen-Cheng Liu, Yi-Chiao Wu, Yue Zhao, Yuval Kirstain, Zecheng He, Zijian He, Albert Pumarola, Ali Thabet, Artsiom Sanakoyeu, Arun Mallya, Baishan Guo, Boris Araya, Breena Kerr, Carleigh Wood, Ce Liu, Cen Peng, Dimitry Vengertsev, Edgar Schonfeld, Elliot Blanchard, Felix Juefei-Xu, Fraylie Nord, Jeff Liang, John Hoffman, Jonas Kohler, Kaolin Fire, Karthik Sivakumar, Lawrence Chen, Licheng Yu, Luya Gao, Markos Georgopoulos, Rashel Moritz, Sara K. Sampson, Shikai Li, Simone Parmeggiani, Steve Fine, Tara Fowler, Vladan Petrovic, and Yuming Du. Movie gen: A cast of media foundation models, 2024. 1" + } + ] + } + ], + "index": 3 + }, + { + "bbox": [ + 56, + 453, + 294, + 518 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 453, + 294, + 518 + ], + "spans": [ + { + "bbox": [ + 56, + 453, + 294, + 518 + ], + "type": "text", + "content": "[32] Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. In International Conference on Machine Learning, pages 8748-8763, 2021. 2, 3, 5" + } + ] + } + ], + "index": 4 + }, + { + "bbox": [ + 56, + 520, + 294, + 563 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 520, + 294, + 563 + ], + "spans": [ + { + "bbox": [ + 56, + 520, + 294, + 563 + ], + "type": "text", + "content": "[33] Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Björn Ommer. High-resolution image synthesis with latent diffusion models. In CVPR, pages 10684-10695, 2022. 1, 3, 5" + } + ] + } + ], + "index": 5 + }, + { + "bbox": [ + 56, + 566, + 294, + 597 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 56, + 566, + 294, + 597 + ], + "spans": [ + { + "bbox": [ + 56, + 566, + 294, + 597 + ], + "type": "text", + "content": "[34] Roy Sheffer and Yossi Adi. I hear your true colors: Image guided audio generation. In ICASSP, pages 1-5, 2023. 1, 2, 7" + } + ] + } + ], + "index": 6 + }, + { + "bbox": [ + 57, + 600, + 294, + 622 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 57, + 600, + 294, + 622 + ], + "spans": [ + { + "bbox": [ + 57, + 600, + 294, + 622 + ], + "type": "text", + "content": "[35] Jiaming Song, Chenlin Meng, and Stefano Ermon. Denoising diffusion implicit models. In ICLR, 2021. 2" + } + ] + } + ], + "index": 7 + }, + { + "bbox": [ + 57, + 624, + 294, + 666 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 57, + 624, + 294, + 666 + ], + "spans": [ + { + "bbox": [ + 57, + 624, + 294, + 666 + ], + "type": "text", + "content": "[36] Jianlin Su, Murtadha Ahmed, Yu Lu, Shengfeng Pan, Wen Bo, and Yunfeng Liu. Roformer: Enhanced transformer with rotary position embedding. Neurocomputing, 568:127063, 2024. 4" + } + ] + } + ], + "index": 8 + }, + { + "bbox": [ + 57, + 670, + 294, + 713 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 57, + 670, + 294, + 713 + ], + "spans": [ + { + "bbox": [ + 57, + 670, + 294, + 713 + ], + "type": "text", + "content": "[37] Kun Su, Xiulong Liu, and Eli Shlizerman. From vision to audio and beyond: A unified model for audio-visual representation and generation. In International Conference on Machine Learning, pages 46804–46822, 2024. 2, 7" + } + ] + } + ], + "index": 9 + } + ], + "sub_type": "ref_text" + }, + { + "bbox": [ + 316, + 72, + 553, + 552 + ], + "type": "list", + "angle": 0, + "index": 21, + "blocks": [ + { + "bbox": [ + 316, + 72, + 553, + 106 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 72, + 553, + 106 + ], + "spans": [ + { + "bbox": [ + 316, + 72, + 553, + 106 + ], + "type": "text", + "content": "[38] Anni Tang, Tianyu He, Junliang Guo, Xinle Cheng, Li Song, and Jiang Bian. Vidtok: A versatile and open-source video tokenizer. arXiv preprint arXiv:2412.13061, 2024. 2, 3, 5" + } + ] + } + ], + "index": 11 + }, + { + "bbox": [ + 317, + 107, + 553, + 162 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 317, + 107, + 553, + 162 + ], + "spans": [ + { + "bbox": [ + 317, + 107, + 553, + 162 + ], + "type": "text", + "content": "[39] Hugo Touvron, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, et al. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288, 2023. 7" + } + ] + } + ], + "index": 12 + }, + { + "bbox": [ + 317, + 163, + 553, + 215 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 317, + 163, + 553, + 215 + ], + "spans": [ + { + "bbox": [ + 317, + 163, + 553, + 215 + ], + "type": "text", + "content": "[40] Heng Wang, Jianbo Ma, Santiago Pascual, Richard Cartwright, and Weidong Cai. V2a-mapper: A lightweight solution for vision-to-audio generation by connecting foundation models. In AAAI, pages 15492-15501, 2024. 1, 2, 7" + } + ] + } + ], + "index": 13 + }, + { + "bbox": [ + 316, + 219, + 553, + 261 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 219, + 553, + 261 + ], + "spans": [ + { + "bbox": [ + 316, + 219, + 553, + 261 + ], + "type": "text", + "content": "[41] Xihua Wang, Yuyue Wang, Yihan Wu, Ruihua Song, Xu Tan, Zehua Chen, Hongteng Xu, and Guodong Sui. Tiva: Time-aligned video-to-audio generation. In ACM MM, pages 573-582, 2024. 1, 2, 3, 6, 7" + } + ] + } + ], + "index": 14 + }, + { + "bbox": [ + 316, + 263, + 553, + 307 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 263, + 553, + 307 + ], + "spans": [ + { + "bbox": [ + 316, + 263, + 553, + 307 + ], + "type": "text", + "content": "[42] Xiang Wang, Shiwei Zhang, Hangjie Yuan, Zhiwu Qing, Biao Gong, Yingya Zhang, Yujun Shen, Changxin Gao, and Nong Sang. A recipe for scaling up text-to-video generation with text-free videos. In CVPR, pages 6572-6582, 2024. 2" + } + ] + } + ], + "index": 15 + }, + { + "bbox": [ + 316, + 308, + 553, + 372 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 308, + 553, + 372 + ], + "spans": [ + { + "bbox": [ + 316, + 308, + 553, + 372 + ], + "type": "text", + "content": "[43] Yongqi Wang, Wenxiang Guo, Rongjie Huang, Jiawei Huang, Zehan Wang, Fuming You, Ruiqi Li, and Zhou Zhao. Frieren: Efficient video-to-audio generation network with rectified flow matching. In Advances in Neural Information Processing Systems, pages 128118-128138, 2024. 1, 2, 3, 6, 7" + } + ] + } + ], + "index": 16 + }, + { + "bbox": [ + 316, + 375, + 553, + 407 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 375, + 553, + 407 + ], + "spans": [ + { + "bbox": [ + 316, + 375, + 553, + 407 + ], + "type": "text", + "content": "[44] Tianxing Wu, Chenyang Si, Yuming Jiang, Ziqi Huang, and Ziwei Liu. Freeinit: Bridging initialization gap in video diffusion models. In ECCV, pages 378-394, 2024. 2" + } + ] + } + ], + "index": 17 + }, + { + "bbox": [ + 316, + 409, + 553, + 453 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 409, + 553, + 453 + ], + "spans": [ + { + "bbox": [ + 316, + 409, + 553, + 453 + ], + "type": "text", + "content": "[45] Yazhou Xing, Yingqing He, Zeyue Tian, Xintao Wang, and Qifeng Chen. Seeing and hearing: Open-domain visual-audio generation with diffusion latent aligners. In CVPR, pages 7151-7161, 2024. 1, 2, 6, 7" + } + ] + } + ], + "index": 18 + }, + { + "bbox": [ + 316, + 454, + 553, + 507 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 454, + 553, + 507 + ], + "spans": [ + { + "bbox": [ + 316, + 454, + 553, + 507 + ], + "type": "text", + "content": "[46] Yiming Zhang, Yicheng Gu, Yanhong Zeng, Zhening Xing, Yuancheng Wang, Zhizheng Wu, and Kai Chen. Foley-crafter: Bring silent videos to life with lifelike and synchronized sounds. arXiv preprint arXiv:2407.01494, 2024. 2, 3, 6, 7" + } + ] + } + ], + "index": 19 + }, + { + "bbox": [ + 316, + 510, + 553, + 552 + ], + "type": "ref_text", + "angle": 0, + "lines": [ + { + "bbox": [ + 316, + 510, + 553, + 552 + ], + "spans": [ + { + "bbox": [ + 316, + 510, + 553, + 552 + ], + "type": "text", + "content": "[47]Zikai Zhou,Shitong Shao,Lichen Bai,Zhiqiang Xu,Bo Han,and Zeke Xie. Golden noise for diffusion models: A learning framework.arXiv preprint arXiv:2411.09502, 2024.2" + } + ] + } + ], + "index": 20 + } + ], + "sub_type": "ref_text" + } + ], + "discarded_blocks": [ + { + "bbox": [ + 294, + 749, + 318, + 757 + ], + "type": "page_number", + "angle": 0, + "lines": [ + { + "bbox": [ + 294, + 749, + 318, + 757 + ], + "spans": [ + { + "bbox": [ + 294, + 749, + 318, + 757 + ], + "type": "text", + "content": "11786" + } + ] + } + ], + "index": 22 + } + ], + "page_size": [ + 612, + 792 + ], + "page_idx": 9 + } + ], + "_backend": "vlm", + "_version_name": "2.6.4" +} \ No newline at end of file