ChatPaper.aiChatPaper

4DGS360:基于单视频的动态物体360°高斯重建

4DGS360: 360° Gaussian Reconstruction of Dynamic Objects from a Single Video

March 23, 2026
作者: Jae Won Jang, Yeonjin Chang, Wonsik Shin, Juhwan Cho, Nojun Kwak
cs.AI

摘要

我们提出4DGS360——一种无需扩散技术的框架,能从单目手持视频实现360度动态物体重建。现有方法因过度依赖二维先验知识,导致初始点过度拟合训练视角中的可见表面,难以重建一致的360度几何模型。4DGS360通过先进的3D原生初始化技术应对这一挑战,有效缓解被遮挡区域的几何模糊性问题。我们研发的3D追踪器AnchorTAP3D以置信度高的2D追踪点为锚点,生成强化的3D点轨迹,既能抑制漂移现象,又能提供保持被遮挡区域几何结构的可靠初始化方案。这种初始化与优化过程的结合,最终生成连贯的360度四维重建结果。我们还推出了iPhone360新型基准数据集,其测试相机视角与训练视角最大间隔达135度,实现了现有数据集无法支持的360度全景评估。实验表明,4DGS360在iPhone360、iPhone和DAVIS数据集上均取得定性与定量评估的双重最优表现。
English
We introduce 4DGS360, a diffusion-free framework for 360^{circ} dynamic object reconstruction from casual monocular video. Existing methods often fail to reconstruct consistent 360^{circ} geometry, as their heavy reliance on 2D-native priors causes initial points to overfit to visible surface in each training view. 4DGS360 addresses this challenge through a advanced 3D-native initialization that mitigates the geometric ambiguity of occluded regions. Our proposed 3D tracker, AnchorTAP3D, produces reinforced 3D point trajectories by leveraging confident 2D track points as anchors, suppressing drift and providing reliable initialization that preserves geometry in occluded regions. This initialization, combined with optimization, yields coherent 360^{circ} 4D reconstructions. We further present iPhone360, a new benchmark where test cameras are placed up to 135^{circ} apart from training views, enabling 360^{circ} evaluation that existing datasets cannot provide. Experiments show that 4DGS360 achieves state-of-the-art performance on the iPhone360, iPhone, and DAVIS datasets, both qualitatively and quantitatively.
PDF91March 27, 2026