ArXiv TLDR

Quadruped Parkour Learning: Sparsely Gated Mixture of Experts with Visual Input

🐦 Tweet
2604.19344

Michael Ziegltrum, Jianhao Jiao, Tianhu Peng, Chengxu Zhou, Dimitrios Kanoulas

cs.RO

TLDR

This paper uses sparsely gated Mixture-of-Experts (MoE) for vision-based quadruped parkour, doubling success rates with better computational efficiency.

Key contributions

  • Investigates sparsely gated MoE architectures for vision-based quadruped parkour.
  • MoE policy doubles successful obstacle traversals on a real Unitree Go2 robot vs. standard MLPs.
  • Achieves superior performance with matched active parameters, showing better computational efficiency.
  • MLPs need 14.3% more computation to match MoE performance, proving MoE's efficiency advantage.

Why it matters

This paper innovatively applies sparsely gated Mixture-of-Experts to vision-based robotic parkour, significantly improving quadruped performance on challenging terrain. It doubles success rates over MLPs with better computational efficiency. This work paves the way for more scalable and capable robotic control policies.

Original Abstract

Robotic parkour provides a compelling benchmark for advancing locomotion over highly challenging terrain, including large discontinuities such as elevated steps. Recent approaches have demonstrated impressive capabilities, including dynamic climbing and jumping, but typically rely on sequential multilayer perceptron (MLP) architectures with densely activated layers. In contrast, sparsely gated mixture-of-experts (MoE) architectures have emerged in the large language model domain as an effective paradigm for improving scalability and performance by activating only a subset of parameters at inference time. In this work, we investigate the application of sparsely gated MoE architectures to vision-based robotic parkour. We compare control policies based on standard MLPs and MoE architectures under a controlled setting where the number of active parameters at inference time is matched. Experimental results on a real Unitree Go2 quadruped robot demonstrate clear performance gains, with the MoE policy achieving double the number of successful trials in traversing large obstacles compared to a standard MLP baseline. We further show that achieving comparable performance with a standard MLP requires scaling its parameter count to match that of the total MoE model, resulting in a 14.3\% increase in computation time. These results highlight that sparsely gated MoE architectures provide a favorable trade-off between performance and computational efficiency, enabling improved scaling of control policies for vision-based robotic parkour. An anonymized link to the codebase is https://osf.io/v2kqj/files/github?view_only=7977dee10c0a44769184498eaba72e44.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.