ArXiv TLDR

SynAgent: Generalizable Cooperative Humanoid Manipulation via Solo-to-Cooperative Agent Synergy

🐦 Tweet
2604.18557

Wei Yao, Haohan Ma, Hongwen Zhang, Yunlian Sun, Liangjun Xing + 4 more

cs.CV

TLDR

SynAgent enables generalizable cooperative humanoid manipulation by synergizing solo-to-cooperative agent skills, outperforming baselines.

Key contributions

  • Leverages Solo-to-Cooperative Agent Synergy for skill transfer from single-agent data.
  • Introduces interaction-preserving retargeting via an Interact Mesh for semantic integrity.
  • Proposes a single-agent pretraining and adaptation paradigm for synergistic behaviors.
  • Develops a trajectory-conditioned generative policy using a conditional VAE.

Why it matters

Cooperative humanoid manipulation is challenging due to data scarcity and coordination issues. SynAgent addresses these by transferring skills from solo to cooperative agents, enabling scalable and generalizable control. It significantly outperforms existing methods, advancing embodied intelligence.

Original Abstract

Controllable cooperative humanoid manipulation is a fundamental yet challenging problem for embodied intelligence, due to severe data scarcity, complexities in multi-agent coordination, and limited generalization across objects. In this paper, we present SynAgent, a unified framework that enables scalable and physically plausible cooperative manipulation by leveraging Solo-to-Cooperative Agent Synergy to transfer skills from single-agent human-object interaction to multi-agent human-object-human scenarios. To maintain semantic integrity during motion transfer, we introduce an interaction-preserving retargeting method based on an Interact Mesh constructed via Delaunay tetrahedralization, which faithfully maintains spatial relationships among humans and objects. Building upon this refined data, we propose a single-agent pretraining and adaptation paradigm that bootstraps synergistic collaborative behaviors from abundant single-human data through decentralized training and multi-agent PPO. Finally, we develop a trajectory-conditioned generative policy using a conditional VAE, trained via multi-teacher distillation from motion imitation priors to achieve stable and controllable object-level trajectory execution. Extensive experiments demonstrate that SynAgent significantly outperforms existing baselines in both cooperative imitation and trajectory-conditioned control, while generalizing across diverse object geometries. Codes and data will be available after publication. Project Page: http://yw0208.github.io/synagent

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.