论文标题
房间中的神经渲染:Amodal 3D理解和免费视图的封闭场景渲染。
Neural Rendering in a Room: Amodal 3D Understanding and Free-Viewpoint Rendering for the Closed Scene Composed of Pre-Captured Objects
论文作者
论文摘要
作为人类,我们可以从任意观点来理解并想象一个熟悉的场景,而这仍然是计算机的巨大挑战。我们特此提出了一种新颖的解决方案,以基于Amodal 3D场景的新范式来模仿这种人类的感知能力,并具有神经渲染的封闭场景。具体来说,我们首先通过离线阶段在封闭场景中学习对物体的先验知识,这有助于在线舞台,以看不见的家具安排来了解房间。在在线阶段,在不同的布局中,我们使用一个整体神经渲染的优化框架来有效地估算正确的3D场景布局并提供现实的免费观看点渲染。为了处理离线和在线阶段之间的域间隙,我们的方法利用了离线培训中数据增强的组成神经渲染技术。合成数据集和真实数据集的实验表明,我们的两阶段设计可以通过大幅度的差距达到强大的3D场景理解,并超越了竞争的方法,我们还表明,我们现实的免费观看点渲染能够启用各种应用程序,包括场景巡回演出和编辑。代码和数据可在项目网页上找到:https://zju3dv.github.io/nr_in_a_room/。
We, as human beings, can understand and picture a familiar scene from arbitrary viewpoints given a single image, whereas this is still a grand challenge for computers. We hereby present a novel solution to mimic such human perception capability based on a new paradigm of amodal 3D scene understanding with neural rendering for a closed scene. Specifically, we first learn the prior knowledge of the objects in a closed scene via an offline stage, which facilitates an online stage to understand the room with unseen furniture arrangement. During the online stage, given a panoramic image of the scene in different layouts, we utilize a holistic neural-rendering-based optimization framework to efficiently estimate the correct 3D scene layout and deliver realistic free-viewpoint rendering. In order to handle the domain gap between the offline and online stage, our method exploits compositional neural rendering techniques for data augmentation in the offline training. The experiments on both synthetic and real datasets demonstrate that our two-stage design achieves robust 3D scene understanding and outperforms competing methods by a large margin, and we also show that our realistic free-viewpoint rendering enables various applications, including scene touring and editing. Code and data are available on the project webpage: https://zju3dv.github.io/nr_in_a_room/.