E2E-Fly: An Integrated Training-to-Deployment System for End-to-End Quadrotor Autonomy
Fangyu Sun, Fanxing Li, Linzuo Zhang, Yu Hu, Renbiao Jin + 3 more
TLDR
E2E-Fly is an integrated system for end-to-end quadrotor autonomy, enabling robust sim-to-real transfer via differentiable physics and a full-stack workflow.
Key contributions
- Presents E2E-Fly, an integrated training-to-deployment system for end-to-end quadrotor autonomy.
- Features a high-performance simulator with differentiable physics learning and reinforcement learning.
- Implements a two-stage validation strategy and comprehensive sim-to-real alignment methods.
- Successfully deploys policies for six end-to-end control tasks on real quadrotor platforms.
Why it matters
This paper introduces E2E-Fly, a crucial step towards robust end-to-end quadrotor autonomy. It tackles the long-standing challenge of sim-to-real transfer by unifying differentiable physics with a complete training-to-deployment workflow. This system enhances reproducibility and accelerates real-world deployment of learning-based policies.
Original Abstract
Training and transferring learning-based policies for quadrotors from simulation to reality remains challenging due to inefficient visual rendering, physical modeling inaccuracies, unmodeled sensor discrepancies, and the absence of a unified platform integrating differentiable physics learning into end-to-end training. While recent work has demonstrated various end-to-end quadrotor control tasks, few systems provide a systematic, zero-shot transfer pipeline, hindering reproducibility and real-world deployment. To bridge this gap, we introduce E2E-Fly, an integrated framework featuring an agile quadrotor platform coupled with a full-stack training, validation, and deployment workflow. The training framework incorporates a high-performance simulator with support for differentiable physics learning and reinforcement learning, alongside structured reward design tailored to common quadrotor tasks. We further introduce a two-stage validation strategy using sim-to-sim transfer and hardware-in-the-loop testing, and deploy policies onto two physical quadrotor platforms via a dedicated low-level control interface and a comprehensive sim-to-real alignment methodology, encompassing system identification, domain randomization, latency compensation, and noise modeling. To the best of our knowledge, this is the first work to systematically unify differentiable physical learning with training, validation, and real-world deployment for quadrotors. Finally, we demonstrate the effectiveness of our framework for training six end-to-end control tasks and deploy them in the real world.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.