Neural Control: Adjoint Learning Through Equilibrium Constraints
Dezhong Tong, Jiawen Wang, Hengyi Zhou, Yinglong Shen, Xiaonan Huang + 1 more
TLDR
Neural Control uses adjoint learning to efficiently optimize physical AI tasks with implicit equilibrium constraints, outperforming gradient-free methods.
Key contributions
- Introduces Neural Control, a boundary-control framework for physical AI tasks with implicit equilibrium.
- Computes memory-efficient proxy gradients using an adjoint formulation, avoiding costly solver unrolling.
- Integrates gradients into receding-horizon MPC for robust control in multi-stable, long-horizon tasks.
- Demonstrates improved performance over gradient-free baselines on DLO manipulation in sim and real robots.
Why it matters
Many physical AI tasks are hard to control due to implicit dynamics and multi-stability. This paper offers a robust, memory-efficient solution for learning and control in such systems. By using adjoint learning and MPC, it enables more effective manipulation of deformable objects, crucial for robotics and manufacturing.
Original Abstract
Many physical AI tasks are governed by implicit equilibrium: an agent actuates a subset of degrees of freedom (boundary DoFs), while the remaining free DoFs settle by minimizing a total potential energy. Even seemingly basic tasks such as bending a deformable linear object (DLO) to a target shape can exhibit strongly nonlinear behavior due to multi-stability: the same boundary conditions may yield multiple equilibrium shapes depending on the actuation trajectory. However, learning and control in such systems is brittle because the actuation-to-configuration map is defined only implicitly, and naive backpropagation through iterative equilibrium solvers is memory- and compute-intensive. We propose Neural Control, a boundary-control framework that computes trajectory-dependent, memory-efficient proxy gradients by differentiating equilibrium conditions via an adjoint formulation, avoiding unrolling of solver iterations. To improve robustness over long horizons, we integrate these sensitivities into a receding-horizon MPC scheme that repeatedly re-anchors optimization to realized equilibria and mitigates basin-switching in multi-stable regimes. We evaluate Neural Control in simulation and on physical robots manipulating DLOs, and show improved performance over gradient-free baselines such as SPSA and CEM.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.