论文标题
场景:RGB-D扫描中预测对象对齐和布局
SceneCAD: Predicting Object Alignments and Layouts in RGB-D Scans
论文作者
论文摘要
我们提出了一种新的方法,可以从商品RGB-D传感器中重建对经过扫描的3D环境的轻质,基于CAD的表示。我们的关键想法是共同优化CAD模型对齐方式以及对扫描场景的布局估算,明确对对象对象和对象之间的相互关联进行建模。由于对象布置和场景布局本质上是耦合的,因此我们表明,共同处理问题有助于产生场景的全球一致表示。对象CAD模型通过建立几何形状之间的密集对应关系来对齐场景,我们引入了一种层次的布局预测方法,以估算场景的角落和边缘的布局平面。为此,我们提出了一个消息传播的图形神经网络,以模拟对象和布局之间的相互关联,并指导了一个全球对象对象对象对象的场景。通过考虑全球场景布局,与最先进的方法相比,我们实现了明显改善的CAD对齐,从SUNCG上的41.83%到58.41%的一致性准确性,分别从50.05%提高到Scannet的50.05%增加到61.24%。由此产生的基于CAD的表示使我们的方法非常适合在内容创建中的应用,例如增强或虚拟现实。
We present a novel approach to reconstructing lightweight, CAD-based representations of scanned 3D environments from commodity RGB-D sensors. Our key idea is to jointly optimize for both CAD model alignments as well as layout estimations of the scanned scene, explicitly modeling inter-relationships between objects-to-objects and objects-to-layout. Since object arrangement and scene layout are intrinsically coupled, we show that treating the problem jointly significantly helps to produce globally-consistent representations of a scene. Object CAD models are aligned to the scene by establishing dense correspondences between geometry, and we introduce a hierarchical layout prediction approach to estimate layout planes from corners and edges of the scene.To this end, we propose a message-passing graph neural network to model the inter-relationships between objects and layout, guiding generation of a globally object alignment in a scene. By considering the global scene layout, we achieve significantly improved CAD alignments compared to state-of-the-art methods, improving from 41.83% to 58.41% alignment accuracy on SUNCG and from 50.05% to 61.24% on ScanNet, respectively. The resulting CAD-based representations makes our method well-suited for applications in content creation such as augmented- or virtual reality.