ChatPaper.aiChatPaper

CAT3D:使用多视角扩散模型在3D中创作任何内容

CAT3D: Create Anything in 3D with Multi-View Diffusion Models

May 16, 2024
作者: Ruiqi Gao, Aleksander Holynski, Philipp Henzler, Arthur Brussee, Ricardo Martin-Brualla, Pratul Srinivasan, Jonathan T. Barron, Ben Poole
cs.AI

摘要

3D重建的进展实现了高质量的3D捕捉,但需要用户收集数百到数千张图像来创建一个3D场景。我们提出了CAT3D,一种通过模拟这个真实世界捕捉过程的多视角扩散模型来创建任何3D物体的方法。给定任意数量的输入图像和一组目标新视点,我们的模型生成高度一致的场景新视图。这些生成的视图可用作强大的3D重建技术的输入,以生成可以实时从任何视点渲染的3D表示。CAT3D可以在一分钟内创建完整的3D场景,并且在单图像和少视角3D场景创建方面优于现有方法。请访问我们的项目页面以查看结果和交互式演示:https://cat3d.github.io。
English
Advances in 3D reconstruction have enabled high-quality 3D capture, but require a user to collect hundreds to thousands of images to create a 3D scene. We present CAT3D, a method for creating anything in 3D by simulating this real-world capture process with a multi-view diffusion model. Given any number of input images and a set of target novel viewpoints, our model generates highly consistent novel views of a scene. These generated views can be used as input to robust 3D reconstruction techniques to produce 3D representations that can be rendered from any viewpoint in real-time. CAT3D can create entire 3D scenes in as little as one minute, and outperforms existing methods for single image and few-view 3D scene creation. See our project page for results and interactive demos at https://cat3d.github.io .

Summary

AI-Generated Summary

PDF492December 15, 2024