ChatPaper.aiChatPaper

PHUMA:基于物理的人形机器人运动数据集

PHUMA: Physically-Grounded Humanoid Locomotion Dataset

October 30, 2025
作者: Kyungmin Lee, Sibeen Kim, Minho Park, Hyunseung Kim, Dongyoon Hwang, Hojoon Lee, Jaegul Choo
cs.AI

摘要

動作模仿是實現仿人機器人運動的一種前景廣闊的方法,能使智能體習得類人行為。現有方法通常依賴AMASS等高質量動作捕捉數據集,但這類數據稀缺且成本高昂,限制了可擴展性與多樣性。近期研究嘗試通過轉化大規模網絡視頻(如Humanoid-X)來擴展數據收集規模,但常伴隨漂浮、穿透和滑步等物理偽影,影響模仿穩定性。為此,我們提出PHUMA——基於物理約束的仿人運動數據集,該數據集在利用大規模人類視頻的同時,通過精細化數據校準與物理約束重定向技術解決物理偽影問題。PHUMA強制關節活動限制、確保足部接地並消除滑步現象,生成兼具大規模與物理可靠性的運動數據。我們在兩種情境下評估PHUMA:(1)對自錄測試視頻中未見過動作的模仿;(2)僅憑骨盆引導的路徑追蹤。兩種實驗中,基於PHUMA訓練的策略均超越Humanoid-X和AMASS,在模仿多樣化運動方面取得顯著提升。項目代碼已開源於https://davian-robotics.github.io/PHUMA。
English
Motion imitation is a promising approach for humanoid locomotion, enabling agents to acquire humanlike behaviors. Existing methods typically rely on high-quality motion capture datasets such as AMASS, but these are scarce and expensive, limiting scalability and diversity. Recent studies attempt to scale data collection by converting large-scale internet videos, exemplified by Humanoid-X. However, they often introduce physical artifacts such as floating, penetration, and foot skating, which hinder stable imitation. In response, we introduce PHUMA, a Physically-grounded HUMAnoid locomotion dataset that leverages human video at scale, while addressing physical artifacts through careful data curation and physics-constrained retargeting. PHUMA enforces joint limits, ensures ground contact, and eliminates foot skating, producing motions that are both large-scale and physically reliable. We evaluated PHUMA in two sets of conditions: (i) imitation of unseen motion from self-recorded test videos and (ii) path following with pelvis-only guidance. In both cases, PHUMA-trained policies outperform Humanoid-X and AMASS, achieving significant gains in imitating diverse motions. The code is available at https://davian-robotics.github.io/PHUMA.
PDF302January 19, 2026