ChatPaper.aiChatPaper

RenderFormer:基于Transformer的三角形网格神经渲染与全局光照

RenderFormer: Transformer-based Neural Rendering of Triangle Meshes with Global Illumination

May 28, 2025
作者: Chong Zeng, Yue Dong, Pieter Peers, Hongzhi Wu, Xin Tong
cs.AI

摘要

我们提出了RenderFormer,一种神经渲染管线,它能够直接从基于三角形的场景表示中渲染出包含完整全局光照效果的图像,且无需针对每个场景进行训练或微调。不同于传统的以物理为中心的渲染方法,我们将渲染过程建模为一个序列到序列的转换任务,即将代表带有反射属性的三角形序列转换为代表像素小块的输出序列。RenderFormer采用两阶段管线:第一阶段为视角无关阶段,负责建模三角形间的光传输;第二阶段为视角相关阶段,在视角无关阶段生成的三角形序列指导下,将代表光线束的标记转换为相应的像素值。这两个阶段均基于Transformer架构构建,并在最小先验约束下进行学习。我们在形状和光传输复杂度各异的场景上对RenderFormer进行了展示与评估。
English
We present RenderFormer, a neural rendering pipeline that directly renders an image from a triangle-based representation of a scene with full global illumination effects and that does not require per-scene training or fine-tuning. Instead of taking a physics-centric approach to rendering, we formulate rendering as a sequence-to-sequence transformation where a sequence of tokens representing triangles with reflectance properties is converted to a sequence of output tokens representing small patches of pixels. RenderFormer follows a two stage pipeline: a view-independent stage that models triangle-to-triangle light transport, and a view-dependent stage that transforms a token representing a bundle of rays to the corresponding pixel values guided by the triangle-sequence from the view-independent stage. Both stages are based on the transformer architecture and are learned with minimal prior constraints. We demonstrate and evaluate RenderFormer on scenes with varying complexity in shape and light transport.

Summary

AI-Generated Summary

PDF333May 29, 2025