ArXiv TLDR

FlowAnchor: Stabilizing the Editing Signal for Inversion-Free Video Editing

🐦 Tweet
2604.22586

Ze Chen, Lan Chen, Yuanhang Li, Qi Mao

cs.CV

TLDR

FlowAnchor stabilizes the editing signal in inversion-free video editing, enabling stable, efficient, and coherent results in complex multi-object and fast-motion scenarios.

Key contributions

  • Proposes FlowAnchor, a training-free framework for stable, efficient, inversion-free video editing.
  • Stabilizes editing signals by anchoring spatial regions and adaptively modulating editing strength.
  • Uses Spatial-aware Attention Refinement for consistent textual guidance and spatial region alignment.
  • Applies Adaptive Magnitude Modulation to preserve sufficient editing strength across frames.

Why it matters

Inversion-free video editing struggles with stability in complex scenes, limiting its practical application. FlowAnchor addresses this by stabilizing the editing signal, leading to more faithful and temporally coherent edits. This advances video manipulation, making it efficient for multi-object and fast-motion scenarios.

Original Abstract

We propose FlowAnchor, a training-free framework for stable and efficient inversion-free, flow-based video editing. Inversion-free editing methods have recently shown impressive efficiency and structure preservation in images by directly steering the sampling trajectory with an editing signal. However, extending this paradigm to videos remains challenging, often failing in multi-object scenes or with increased frame counts. We identify the root cause as the instability of the editing signal in high-dimensional video latent spaces, which arises from imprecise spatial localization and length-induced magnitude attenuation. To overcome this challenge, FlowAnchor explicitly anchors both where to edit and how strongly to edit. It introduces Spatial-aware Attention Refinement, which enforces consistent alignment between textual guidance and spatial regions, and Adaptive Magnitude Modulation, which adaptively preserves sufficient editing strength. Together, these mechanisms stabilize the editing signal and guide the flow-based evolution toward the desired target distribution. Extensive experiments demonstrate that FlowAnchor achieves more faithful, temporally coherent, and computationally efficient video editing across challenging multi-object and fast-motion scenarios. The project page is available at https://cuc-mipg.github.io/FlowAnchor.github.io/.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.