ChatPaper.aiChatPaper

PHUMA:基于物理的人形机器人运动数据集

PHUMA: Physically-Grounded Humanoid Locomotion Dataset

October 30, 2025
作者: Kyungmin Lee, Sibeen Kim, Minho Park, Hyunseung Kim, Dongyoon Hwang, Hojoon Lee, Jaegul Choo
cs.AI

摘要

运动模仿是实现仿人机器人运动的一种前景广阔的方法,可使智能体获得类人行为。现有方法通常依赖AMASS等高质量运动捕捉数据集,但这些数据稀缺且昂贵,限制了方法的可扩展性与多样性。近期研究尝试通过转换大规模网络视频(如Humanoid-X)来扩大数据采集规模,但常引入漂浮、穿透、足部滑移等物理伪影,阻碍了稳定模仿。为此,我们提出PHUMA——基于物理约束的仿人运动数据集,该方案在利用大规模人类视频数据的同时,通过精细数据筛选与物理约束的重定向技术解决物理伪影问题。PHUMA通过强制关节限制、确保地面接触并消除足部滑移,生成兼具大规模与物理可靠性的运动数据。我们在两种条件下评估PHUMA:(1)对自录测试视频中未见运动的模仿;(2)仅基于骨盆引导的路径跟随。两种场景下,基于PHUMA训练的策略均优于Humanoid-X和AMASS,在多样化运动模仿方面取得显著提升。代码已开源:https://davian-robotics.github.io/PHUMA。
English
Motion imitation is a promising approach for humanoid locomotion, enabling agents to acquire humanlike behaviors. Existing methods typically rely on high-quality motion capture datasets such as AMASS, but these are scarce and expensive, limiting scalability and diversity. Recent studies attempt to scale data collection by converting large-scale internet videos, exemplified by Humanoid-X. However, they often introduce physical artifacts such as floating, penetration, and foot skating, which hinder stable imitation. In response, we introduce PHUMA, a Physically-grounded HUMAnoid locomotion dataset that leverages human video at scale, while addressing physical artifacts through careful data curation and physics-constrained retargeting. PHUMA enforces joint limits, ensures ground contact, and eliminates foot skating, producing motions that are both large-scale and physically reliable. We evaluated PHUMA in two sets of conditions: (i) imitation of unseen motion from self-recorded test videos and (ii) path following with pelvis-only guidance. In both cases, PHUMA-trained policies outperform Humanoid-X and AMASS, achieving significant gains in imitating diverse motions. The code is available at https://davian-robotics.github.io/PHUMA.
PDF302January 19, 2026