Papers
arxiv:2603.29092

TrajectoryMover: Generative Movement of Object Trajectories in Videos

Published on Mar 31
· Submitted by
Kiran Chhatre
on Apr 1
Authors:
,
,
,
,

Abstract

TrajectoryAtlas enables generative video editing by generating large-scale synthetic paired video data and training a video generator to move object 3D motion trajectories while preserving plausibility and identity.

AI-generated summary

Generative video editing has enabled several intuitive editing operations for short video clips that would previously have been difficult to achieve, especially for non-expert editors. Existing methods focus on prescribing an object's 3D or 2D motion trajectory in a video, or on altering the appearance of an object or a scene, while preserving both the video's plausibility and identity. Yet a method to move an object's 3D motion trajectory in a video, i.e., moving an object while preserving its relative 3D motion, is currently still missing. The main challenge lies in obtaining paired video data for this scenario. Previous methods typically rely on clever data generation approaches to construct plausible paired data from unpaired videos, but this approach fails if one of the videos in a pair can not easily be constructed from the other. Instead, we introduce TrajectoryAtlas, a new data generation pipeline for large-scale synthetic paired video data and a video generator TrajectoryMover fine-tuned with this data. We show that this successfully enables generative movement of object trajectories. Project page: https://chhatrekiran.github.io/trajectorymover

Community

Paper author Paper submitter

TrajectoryMover is a video-to-video editing method that moves an object’s trajectory while preserving its relative 3D motion, enabled by a large-scale synthetic paired video data generation pipeline for this task.

Sign up or log in to comment

Get this paper in your agent:

hf papers read 2603.29092
Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2603.29092 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2603.29092 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2603.29092 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.