ChatPaper.aiChatPaper

ObjectGS:基於高斯潑濺的物件感知場景重建與場景理解

ObjectGS: Object-aware Scene Reconstruction and Scene Understanding via Gaussian Splatting

July 21, 2025
作者: Ruijie Zhu, Mulin Yu, Linning Xu, Lihan Jiang, Yixuan Li, Tianzhu Zhang, Jiangmiao Pang, Bo Dai
cs.AI

摘要

3D高斯潑濺技術以其高保真重建與實時新視角合成而著稱,然而其缺乏語義理解能力,限制了物體層面的感知。在本研究中,我們提出了ObjectGS,這是一個物體感知框架,將3D場景重建與語義理解相統一。ObjectGS不再將場景視為一個整體,而是將各個物體建模為生成神經高斯並共享物體ID的局部錨點,從而實現精確的物體層面重建。在訓練過程中,我們動態地增長或修剪這些錨點,並優化其特徵,同時利用帶有分類損失的獨熱ID編碼來強化清晰的語義約束。通過大量實驗,我們展示了ObjectGS不僅在開放詞彙和全景分割任務上超越了現有最先進的方法,還能無縫集成於網格提取和場景編輯等應用中。項目頁面:https://ruijiezhu94.github.io/ObjectGS_page
English
3D Gaussian Splatting is renowned for its high-fidelity reconstructions and real-time novel view synthesis, yet its lack of semantic understanding limits object-level perception. In this work, we propose ObjectGS, an object-aware framework that unifies 3D scene reconstruction with semantic understanding. Instead of treating the scene as a unified whole, ObjectGS models individual objects as local anchors that generate neural Gaussians and share object IDs, enabling precise object-level reconstruction. During training, we dynamically grow or prune these anchors and optimize their features, while a one-hot ID encoding with a classification loss enforces clear semantic constraints. We show through extensive experiments that ObjectGS not only outperforms state-of-the-art methods on open-vocabulary and panoptic segmentation tasks, but also integrates seamlessly with applications like mesh extraction and scene editing. Project page: https://ruijiezhu94.github.io/ObjectGS_page
PDF71July 23, 2025