Tree Learning: A Multi-Skill Continual Learning Framework for Humanoid Robots
TLDR
Tree Learning is a multi-skill continual learning framework for humanoid robots that prevents catastrophic forgetting and enables efficient skill expansion.
Key contributions
- Root-branch hierarchical parameter inheritance prevents catastrophic forgetting via motion priors.
- Multi-modal feedforward adaptation supports periodic/aperiodic motions using phase modulation & interpolation.
- Task-level reward shaping accelerates skill convergence for new robot tasks.
- Achieves higher rewards and 100% skill retention in simulations over multi-task training.
Why it matters
This paper addresses a critical challenge in embodied intelligence: enabling humanoid robots to continually learn new skills without forgetting old ones. Tree Learning offers a lightweight, efficient solution that ensures high performance and skill retention, paving the way for more versatile and adaptable robots.
Original Abstract
As reinforcement learning for humanoid robots evolves from single-task to multi-skill paradigms, efficiently expanding new skills while avoiding catastrophic forgetting has become a key challenge in embodied intelligence. Existing approaches either rely on complex topology adjustments in Mixture-of-Experts (MoE) models or require training extremely large-scale models, making lightweight deployment difficult. To address this, we propose Tree Learning, a multi-skill continual learning framework for humanoid robots. The framework adopts a root-branch hierarchical parameter inheritance mechanism, providing motion priors for branch skills through parameter reuse to fundamentally prevent catastrophic forgetting. A multi-modal feedforward adaptation mechanism combining phase modulation and interpolation is designed to support both periodic and aperiodic motions. A task-level reward shaping strategy is also proposed to accelerate skill convergence. Unity-based simulation experiments show that, in contrast to simultaneous multi-task training, Tree Learning achieves higher rewards across various representative locomotion skills while maintaining a 100% skill retention rate, enabling seamless multi-skill switching and real-time interactive control. We further validate the performance and generalization capability of Tree Learning on two distinct Unity-simulated tasks: a Super Mario-inspired interactive scenario and autonomous navigation in a classical Chinese garden environment.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.