论文标题
几何对应场:在野外进行3D姿势改进的可区分渲染
Geometric Correspondence Fields: Learned Differentiable Rendering for 3D Pose Refinement in the Wild
论文作者
论文摘要
我们提出了一种基于野外任意类别对象的可区分渲染的新颖3D姿势完善方法。与以前的方法相反,我们做出了两个主要贡献:首先,不是在RGB或蒙版空间中比较现实世界的图像和合成渲染,而是在优化3D姿势细化的特征空间中进行比较。其次,我们介绍了一个新颖的可区分渲染器,该渲染器学会了从数据向后近似栅格化,而不是依靠手工制作的算法。为此,我们预测了RGB图像与3D模型渲染之间的深层跨域对应关系,以我们称为几何对应字段的形式。这些对应字段用作像素级梯度,通过渲染管道在分析上向后传播,以直接在3D姿势上进行基于梯度的优化。通过这种方式,我们将3D模型与RGB图像中的对象完全归结为,从而显着改善了3D姿势估计。我们在具有挑战性的Pix3D数据集上评估了我们的方法,并且与多个指标的最新精炼方法相比,相对改进的相对改进高达55%。
We present a novel 3D pose refinement approach based on differentiable rendering for objects of arbitrary categories in the wild. In contrast to previous methods, we make two main contributions: First, instead of comparing real-world images and synthetic renderings in the RGB or mask space, we compare them in a feature space optimized for 3D pose refinement. Second, we introduce a novel differentiable renderer that learns to approximate the rasterization backward pass from data instead of relying on a hand-crafted algorithm. For this purpose, we predict deep cross-domain correspondences between RGB images and 3D model renderings in the form of what we call geometric correspondence fields. These correspondence fields serve as pixel-level gradients which are analytically propagated backward through the rendering pipeline to perform a gradient-based optimization directly on the 3D pose. In this way, we precisely align 3D models to objects in RGB images which results in significantly improved 3D pose estimates. We evaluate our approach on the challenging Pix3D dataset and achieve up to 55% relative improvement compared to state-of-the-art refinement methods in multiple metrics.