World-R1: Reinforcing 3D Constraints for Text-to-Video Generation
Weijie Wang, Xiaoxuan He, Youping Gu, Yifan Yang, Zeyu Zhang + 7 more
TLDR
World-R1 uses reinforcement learning to enforce 3D constraints in text-to-video generation, improving geometric consistency without architectural changes.
Key contributions
- Aligns video generation with 3D constraints via reinforcement learning.
- Introduces a specialized pure text dataset tailored for world simulation.
- Optimizes with Flow-GRPO using feedback from pre-trained 3D foundation and VL models.
- Employs periodic decoupled training for rigid geometric consistency and dynamic scene fluidity.
Why it matters
This paper addresses geometric inconsistencies in text-to-video models without costly architectural changes. By using RL and a novel training strategy, World-R1 significantly enhances 3D consistency, paving the way for more realistic and scalable world simulations.
Original Abstract
Recent video foundation models demonstrate impressive visual synthesis but frequently suffer from geometric inconsistencies. While existing methods attempt to inject 3D priors via architectural modifications, they often incur high computational costs and limit scalability. We propose World-R1, a framework that aligns video generation with 3D constraints through reinforcement learning. To facilitate this alignment, we introduce a specialized pure text dataset tailored for world simulation. Utilizing Flow-GRPO, we optimize the model using feedback from pre-trained 3D foundation models and vision-language models to enforce structural coherence without altering the underlying architecture. We further employ a periodic decoupled training strategy to balance rigid geometric consistency with dynamic scene fluidity. Extensive evaluations reveal that our approach significantly enhances 3D consistency while preserving the original visual quality of the foundation model, effectively bridging the gap between video generation and scalable world simulation.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.