ArXiv TLDR

MixFlow: Mixed Source Distributions Improve Rectified Flows

🐦 Tweet
2604.09181

Nazir Nayal, Christopher Wewer, Jan Eric Lenssen

cs.CVcs.LG

TLDR

MixFlow improves rectified flows by mixing source distributions, leading to faster, higher-quality image generation with reduced path curvature.

Key contributions

  • Introduces κ-FC, a general formulation to condition source distributions on arbitrary signals.
  • Presents MixFlow, a training strategy that reduces generative path curvatures.
  • MixFlow trains on linear mixtures of unconditional and κ-FC-based distributions.
  • Achieves 12% FID improvement and faster training convergence for rectified flows.

Why it matters

Diffusion models are hindered by slow sampling due to highly curved generative paths. MixFlow addresses this by better aligning source and data distributions, significantly improving sampling efficiency and image quality. This advancement makes high-quality generative models more practical and accessible.

Original Abstract

Diffusion models and their variations, such as rectified flows, generate diverse and high-quality images, but they are still hindered by slow iterative sampling caused by the highly curved generative paths they learn. An important cause of high curvature, as shown by previous work, is independence between the source distribution (standard Gaussian) and the data distribution. In this work, we tackle this limitation by two complementary contributions. First, we attempt to break away from the standard Gaussian assumption by introducing $κ\texttt{-FC}$, a general formulation that conditions the source distribution on an arbitrary signal $κ$ that aligns it better with the data distribution. Then, we present MixFlow, a simple but effective training strategy that reduces the generative path curvatures and considerably improves sampling efficiency. MixFlow trains a flow model on linear mixtures of a fixed unconditional distribution and a $κ\texttt{-FC}$-based distribution. This simple mixture improves the alignment between the source and data, provides better generation quality with less required sampling steps, and accelerates the training convergence considerably. On average, our training procedure improves the generation quality by 12\% in FID compared to standard rectified flow and 7\% compared to previous baselines under a fixed sampling budget. Code available at: $\href{https://github.com/NazirNayal8/MixFlow}{https://github.com/NazirNayal8/MixFlow}$

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.