SMAP: Self-supervised Motion Adaptation for Physically Plausible Humanoid Whole-body Control

1Wuhan University, 2Shanghai AI Lab, 3Harbin Institute of Technology (Shenzhen), 4The Chinese University of Hong Kong, 5Fudan University, 6National University of Singapore
-->

Abstract

This paper presents a novel framework that enables real-world humanoid robots to maintain stability while performing human-like motion. Current methods train a policy which allows humanoid robots to follow human body using the massive retargeted human data via reinforcement learning. However, due to the heterogeneity between human and humanoid robot motion, directly using retargeted human motion reduces training efficiency and stability. To this end, we introduce SMAP, a novel whole-body tracking framework that bridges the gap between human and humanoid action spaces, enabling accurate motion mimicry by humanoid robots. The core idea is to use a vector-quantized periodic autoencoder to capture generic atomic behaviors and adapt human motion into physically plausible humanoid motion. This adaptation accelerates training convergence and improves stability when handling novel or challenging motions. We then employ a privileged teacher to distill precise mimicry skills into the student policy with a proposed decoupled reward. We conduct experiments in simulation and real world and demonstrate the superiority stability and performance of our approach over state-of-the-art methods, offering practical guidelines for advancing whole-body control in humanoid robots.

Pipeline
Overview of the proposed pipeline.

Whole-body Control in H1

Whole-body Control in Isaac Gym