ChatPaper.aiChatPaper

CAT3D:使用多視圖擴散模型在3D中創建任何物件

CAT3D: Create Anything in 3D with Multi-View Diffusion Models

May 16, 2024
作者: Ruiqi Gao, Aleksander Holynski, Philipp Henzler, Arthur Brussee, Ricardo Martin-Brualla, Pratul Srinivasan, Jonathan T. Barron, Ben Poole
cs.AI

摘要

3D重建技術的進步已經實現高質量的3D捕捉,但需要用戶收集數百至數千張圖像來創建3D場景。我們提出了CAT3D,一種通過模擬這種現實世界捕捉過程的多視圖擴散模型來創建3D中的任何物體的方法。給定任意數量的輸入圖像和一組目標新視角,我們的模型生成高度一致的場景新視角。這些生成的視角可以作為強大的3D重建技術的輸入,以產生可以從任何視角實時渲染的3D表示。CAT3D可以在一分鐘內創建完整的3D場景,並且優於現有的單圖像和少視角3D場景創建方法。請查看我們的項目頁面以獲得結果和互動演示:https://cat3d.github.io。
English
Advances in 3D reconstruction have enabled high-quality 3D capture, but require a user to collect hundreds to thousands of images to create a 3D scene. We present CAT3D, a method for creating anything in 3D by simulating this real-world capture process with a multi-view diffusion model. Given any number of input images and a set of target novel viewpoints, our model generates highly consistent novel views of a scene. These generated views can be used as input to robust 3D reconstruction techniques to produce 3D representations that can be rendered from any viewpoint in real-time. CAT3D can create entire 3D scenes in as little as one minute, and outperforms existing methods for single image and few-view 3D scene creation. See our project page for results and interactive demos at https://cat3d.github.io .

Summary

AI-Generated Summary

PDF492December 15, 2024